2025-04-02 03:33:29,086 [ 259455 ] INFO : ClickHouse root is not set. Will use /home/ubuntu/_work/ClickHouse/ClickHouse (runner:53, check_args_and_update_paths) 2025-04-02 03:33:29,086 [ 259455 ] INFO : Cases dir is not set. Will use /home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration (runner:97, check_args_and_update_paths) 2025-04-02 03:33:29,086 [ 259455 ] INFO : utils dir is not set. Will use /home/ubuntu/_work/ClickHouse/ClickHouse/utils (runner:108, check_args_and_update_paths) 2025-04-02 03:33:29,086 [ 259455 ] INFO : base_configs_dir: /home/ubuntu/_work/ClickHouse/ClickHouse/programs/server, binary: /home/ubuntu/_work/_temp/test/build/clickhouse, cases_dir: /home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration (runner:110, check_args_and_update_paths) clickhouse_integration_tests_volume Running pytest container as: 'docker run --rm --name clickhouse_integration_tests_q2ewey --privileged --dns-search='.' --memory=30709035008 --security-opt seccomp=unconfined --cap-add=SYS_PTRACE --volume=/home/ubuntu/_work/_temp/test/build/clickhouse-odbc-bridge:/clickhouse-odbc-bridge --volume=/home/ubuntu/_work/_temp/test/build/clickhouse:/clickhouse --volume=/home/ubuntu/_work/_temp/test/build/clickhouse-library-bridge:/clickhouse-library-bridge --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/programs/server:/clickhouse-config --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration:/ClickHouse/tests/integration --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/utils/backupview:/ClickHouse/utils/backupview --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/utils/grpc-client/pb2:/ClickHouse/utils/grpc-client/pb2 --volume=/run:/run/host:ro --volume=clickhouse_integration_tests_volume:/var/lib/docker -e DOCKER_DOTNET_CLIENT_TAG=11de0b29a15d -e DOCKER_HELPER_TAG=5dc43a6382f0 -e DOCKER_BASE_TAG=8b2301119731 -e DOCKER_KERBEROS_KDC_TAG=9391ecdee8d7 -e DOCKER_MYSQL_GOLANG_CLIENT_TAG=9bec2a638e6e -e DOCKER_MYSQL_JAVA_CLIENT_TAG=766bff31cfe4 -e DOCKER_MYSQL_JS_CLIENT_TAG=41ba7c2ec2a1 -e DOCKER_MYSQL_PHP_CLIENT_TAG=88be89c1e3b6 -e DOCKER_NGINX_DAV_TAG=b55ac9cd7519 -e DOCKER_POSTGRESQL_JAVA_CLIENT_TAG=a4eff5c7f4d6 -e DOCKER_PYTHON_BOTTLE_TAG=caad4729259e -e DOCKER_CLIENT_TIMEOUT=300 -e COMPOSE_HTTP_TIMEOUT=600 -e PYTHONUNBUFFERED=1 -e PYTEST_ADDOPTS="--dist=loadfile -n 10 -rfEps --run-id=0 --color=no --durations=0 test_access_for_functions/test.py::test_access_rights_for_function test_access_for_functions/test.py::test_ignore_obsolete_grant_on_database test_aggregation_memory_efficient/test.py::test_remote test_allowed_client_hosts/test.py::test_allowed_host test_allowed_client_hosts/test.py::test_denied_host test_alter_codec/test.py::test_alter_codec_index test_alter_codec/test.py::test_alter_codec_pk test_alter_on_mixed_type_cluster/test.py::test_alter_on_cluter_non_replicated test_alter_on_mixed_type_cluster/test.py::test_alter_replicated_on_cluster test_alternative_keeper_config/test.py::test_create_insert test_async_connect_to_multiple_ips/test.py::test test_async_metrics_in_cgroup/test.py::test_normalized_user_cpu test_async_metrics_in_cgroup/test.py::test_system_wide_metrics test_async_metrics_in_cgroup/test.py::test_user_cpu_accounting 'test_attach_partition_with_large_destination/test.py::test_attach_partition_with_large_destination[Atomic]' 'test_attach_partition_with_large_destination/test.py::test_attach_partition_with_large_destination[Ordinary]' test_attach_table_normalizer/test.py::test_attach_substr test_attach_table_normalizer/test.py::test_attach_substr_restart test_authentication/test.py::test_authentication_fail test_authentication/test.py::test_authentication_pass test_azure_blob_storage_zero_copy_replication/test.py::test_zero_copy_replication test_backup_restore/test.py::test_attach_partition test_backup_restore/test.py::test_replace_partition test_backup_restore/test.py::test_restore test_backup_restore_new/test_shutdown_wait_backup.py::test_shutdown_wait_backup 'test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-False]' 'test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-True]' 'test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[native-True]' test_backup_restore_on_cluster/test.py::test_backup_restore_on_single_replica test_backup_restore_on_cluster/test.py::test_different_tables_on_nodes test_backup_restore_on_cluster/test.py::test_empty_replicated_table test_backup_restore_on_cluster/test.py::test_file_deduplication test_backup_restore_on_cluster/test.py::test_get_error_from_other_host test_backup_restore_on_cluster/test.py::test_keeper_value_max_size test_backup_restore_on_cluster/test.py::test_mutation test_backup_restore_on_cluster/test.py::test_projection test_backup_restore_on_cluster/test.py::test_replicated_database test_backup_restore_on_cluster/test.py::test_replicated_database_async test_backup_restore_on_cluster/test.py::test_replicated_database_compare_parts 'test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[database]' 'test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[uuid]' test_backup_restore_on_cluster/test.py::test_replicated_table test_backup_restore_on_cluster/test.py::test_replicated_table_after_alters test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_bigger_cluster test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_smaller_cluster test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_def test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_insert test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_merge test_backup_restore_on_cluster/test.py::test_replicated_table_with_uuid_in_zkpath test_backup_restore_on_cluster/test.py::test_required_privileges test_backup_restore_on_cluster/test.py::test_shutdown_waits_for_backup test_backup_restore_on_cluster/test.py::test_system_functions test_backup_restore_on_cluster/test.py::test_system_users test_backup_restore_on_cluster/test.py::test_table_in_replicated_database_with_not_synced_def test_backup_restore_on_cluster/test.py::test_table_with_parts_in_queue_considered_non_empty test_backup_restore_on_cluster/test.py::test_tables_dependency test_backward_compatibility/test_convert_ordinary.py::test_convert_ordinary_to_atomic test_backward_compatibility/test_cte_distributed.py::test_cte_distributed test_backward_compatibility/test_functions.py::test_aggregate_states test_backward_compatibility/test_functions.py::test_string_functions test_backward_compatibility/test_insert_profile_events.py::test_new_client_compatible test_backward_compatibility/test_insert_profile_events.py::test_old_client_compatible test_backward_compatibility/test_normalized_count_comparison.py::test_select_aggregate_alias_column test_backward_compatibility/test_select_aggregate_alias_column.py::test_select_aggregate_alias_column test_backward_compatibility/test_vertical_merges_from_compact_parts.py::test_vertical_merges_from_compact_parts test_broken_part_during_merge/test.py::test_merge_and_part_corruption test_buffer_profile/test.py::test_buffer_profile test_buffer_profile/test.py::test_default_profile test_build_sets_from_multiple_threads/test.py::test_set test_cancel_freeze/test.py::test_cancel_backup test_cgroup_limit/test.py::test_cgroup_cpu_limit test_check_table/test.py::test_check_all_tables 'test_check_table/test.py::test_check_normal_table_corruption[]' 'test_check_table/test.py::test_check_replicated_table_corruption[-_0-.bin]' 'test_check_table/test.py::test_check_replicated_table_simple[-_0]' test_cleanup_after_start/test.py::test_old_dirs_cleanup test_cluster_all_replicas/test.py::test_cluster 'test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[one_shard_three_nodes]' 'test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[two_shards_three_nodes]' test_cluster_all_replicas/test.py::test_global_in 'test_cluster_all_replicas/test.py::test_skip_unavailable_replica[one_shard_three_nodes]' 'test_cluster_all_replicas/test.py::test_skip_unavailable_replica[two_shards_three_nodes]' test_cluster_discovery/test_password.py::test_connect_with_password test_compatibility_merge_tree_settings/test.py::test_check_projections_compatibility test_compatibility_merge_tree_settings/test.py::test_config_overrides_compatibility test_compression_nested_columns/test.py::test_nested_compression_codec test_concurrent_queries_for_all_users_restriction/test.py::test_exception_message test_concurrent_threads_soft_limit/test.py::test_concurrent_threads_soft_limit_default test_concurrent_threads_soft_limit/test.py::test_concurrent_threads_soft_limit_defined_1 test_concurrent_threads_soft_limit/test.py::test_concurrent_threads_soft_limit_defined_50 test_concurrent_threads_soft_limit/test.py::test_concurrent_threads_soft_limit_limit_reached test_concurrent_threads_soft_limit/test.py::test_use_concurrency_control_default test_concurrent_threads_soft_limit/test.py::test_use_concurrency_control_soft_limit_defined_50 test_config_substitutions/test.py::test_allow_databases test_config_substitutions/test.py::test_config test_config_substitutions/test.py::test_config_from_env_overrides test_config_substitutions/test.py::test_config_merge_from_env_overrides test_config_substitutions/test.py::test_config_multiple_zk_substitutions test_config_substitutions/test.py::test_include_config test_custom_settings/test.py::test_custom_settings -vvv -ss" altinityinfra/integration-tests-runner:2165613c5fcd '. Start tests ============================= test session starts ============================== platform linux -- Python 3.10.12, pytest-7.4.4, pluggy-1.5.0 -- /usr/bin/python3 cachedir: .pytest_cache Test order randomisation NOT enabled. Enable with --random-order or --random-order-bucket= rootdir: /ClickHouse/tests/integration configfile: pytest.ini plugins: timeout-2.3.1, repeat-0.9.3, order-1.0.0, reportlog-0.4.0, xdist-3.5.0, random-order-1.1.1 timeout: 900.0s timeout method: signal timeout func_only: False created: 10/10 workers 10 workers [100 items] scheduling tests via LoadFileScheduling Command:[docker ps | wc -l] test_cluster_all_replicas/test.py::test_cluster test_config_substitutions/test.py::test_allow_databases test_check_table/test.py::test_check_all_tables Command:[docker ps | wc -l] Command:[docker ps | wc -l] Command:[docker ps | wc -l] Command:[docker ps | wc -l] Command:[docker ps | wc -l] Command:[docker ps | wc -l] Command:[docker ps | wc -l] test_backup_restore/test.py::test_attach_partition test_alter_codec/test.py::test_alter_codec_index test_allowed_client_hosts/test.py::test_allowed_host test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-False] test_access_for_functions/test.py::test_access_rights_for_function Command:[docker ps | wc -l] Command:[docker ps | wc -l] test_concurrent_threads_soft_limit/test.py::test_concurrent_threads_soft_limit_default test_async_metrics_in_cgroup/test.py::test_normalized_user_cpu Stdout:1 Stdout:1 Stdout:1 No running containers No running containers Pruning Docker networks No running containers Command:[docker network prune --force] Pruning Docker networks Pruning Docker networks Command:[docker network prune --force] Command:[docker network prune --force] Stdout:1 No running containers Pruning Docker networks Command:[docker network prune --force] Stdout:1 No running containers Pruning Docker networks Command:[docker network prune --force] Stdout:1 No running containers Pruning Docker networks Command:[docker network prune --force] Stdout:1 No running containers Pruning Docker networks Command:[docker network prune --force] Stdout:1 No running containers Pruning Docker networks Command:[docker network prune --force] Stdout:1 No running containers Pruning Docker networks Command:[docker network prune --force] Stdout:1 No running containers Pruning Docker networks Command:[docker network prune --force] Command:[sysctl net.ipv4.ip_local_port_range='55000 65535'] Command:[sysctl net.ipv4.ip_local_port_range='55000 65535'] Stdout:net.ipv4.ip_local_port_range = 55000 65535 Command:[sysctl net.ipv4.ip_local_port_range='55000 65535'] Command:[sysctl net.ipv4.ip_local_port_range='55000 65535'] Command:[sysctl net.ipv4.ip_local_port_range='55000 65535'] Running tests in /ClickHouse/tests/integration/test_async_metrics_in_cgroup/test.py Command:[sysctl net.ipv4.ip_local_port_range='55000 65535'] Command:[sysctl net.ipv4.ip_local_port_range='55000 65535'] Cluster start called. is_up=False Stdout:net.ipv4.ip_local_port_range = 55000 65535 Running tests in /ClickHouse/tests/integration/test_allowed_client_hosts/test.py Cluster start called. is_up=False Stdout:net.ipv4.ip_local_port_range = 55000 65535 Stdout:net.ipv4.ip_local_port_range = 55000 65535 Running tests in /ClickHouse/tests/integration/test_config_substitutions/test.py Stdout:net.ipv4.ip_local_port_range = 55000 65535 Cluster start called. is_up=False Running tests in /ClickHouse/tests/integration/test_concurrent_threads_soft_limit/test.py Stdout:net.ipv4.ip_local_port_range = 55000 65535 Running tests in /ClickHouse/tests/integration/test_backup_restore/test.py Cluster start called. is_up=False Cluster start called. is_up=False Running tests in /ClickHouse/tests/integration/test_cluster_all_replicas/test.py Cluster start called. is_up=False Stdout:net.ipv4.ip_local_port_range = 55000 65535 Running tests in /ClickHouse/tests/integration/test_backup_restore_on_cluster/test.py Cluster start called. is_up=False Command:[sysctl net.ipv4.ip_local_port_range='55000 65535'] Command:[sysctl net.ipv4.ip_local_port_range='55000 65535'] Stdout:net.ipv4.ip_local_port_range = 55000 65535 Command:[sysctl net.ipv4.ip_local_port_range='55000 65535'] Running tests in /ClickHouse/tests/integration/test_access_for_functions/test.py Cluster start called. is_up=False Stdout:net.ipv4.ip_local_port_range = 55000 65535 Running tests in /ClickHouse/tests/integration/test_alter_codec/test.py Cluster start called. is_up=False Stdout:net.ipv4.ip_local_port_range = 55000 65535 Running tests in /ClickHouse/tests/integration/test_check_table/test.py Cluster start called. is_up=False Docker networks for project roottestallowedclienthosts-gw8 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestbackuprestore-gw5 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestasyncmetricsincgroup-gw6 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestbackuprestoreoncluster-gw0 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestconfigsubstitutions-gw3 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestclusterallreplicas-gw1 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestaltercodec-gw9 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestaccessforfunctions-gw7 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestchecktable-gw4 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestconcurrentthreadssoftlimit-gw2 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestasyncmetricsincgroup-gw6 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestallowedclienthosts-gw8 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestconfigsubstitutions-gw3 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestbackuprestore-gw5 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestbackuprestoreoncluster-gw0 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestchecktable-gw4 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestaccessforfunctions-gw7 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestclusterallreplicas-gw1 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestaltercodec-gw9 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestconcurrentthreadssoftlimit-gw2 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestallowedclienthosts-gw8 are DRIVER VOLUME NAME Cleanup called Docker volumes for project roottestconfigsubstitutions-gw3 are DRIVER VOLUME NAME Docker volumes for project roottestasyncmetricsincgroup-gw6 are DRIVER VOLUME NAME Cleanup called Cleanup called Docker volumes for project roottestbackuprestoreoncluster-gw0 are DRIVER VOLUME NAME Cleanup called Docker volumes for project roottestchecktable-gw4 are DRIVER VOLUME NAME Cleanup called Docker volumes for project roottestaccessforfunctions-gw7 are DRIVER VOLUME NAME Cleanup called Docker volumes for project roottestclusterallreplicas-gw1 are DRIVER VOLUME NAME Cleanup called Docker volumes for project roottestbackuprestore-gw5 are DRIVER VOLUME NAME Cleanup called Docker volumes for project roottestaltercodec-gw9 are DRIVER VOLUME NAME Cleanup called Docker volumes for project roottestconcurrentthreadssoftlimit-gw2 are DRIVER VOLUME NAME Cleanup called Docker networks for project roottestallowedclienthosts-gw8 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestconfigsubstitutions-gw3 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestasyncmetricsincgroup-gw6 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestchecktable-gw4 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestbackuprestoreoncluster-gw0 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestaccessforfunctions-gw7 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestconcurrentthreadssoftlimit-gw2 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestbackuprestore-gw5 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestaltercodec-gw9 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestclusterallreplicas-gw1 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestconfigsubstitutions-gw3 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestallowedclienthosts-gw8 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestasyncmetricsincgroup-gw6 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestbackuprestoreoncluster-gw0 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestaccessforfunctions-gw7 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestbackuprestore-gw5 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestconcurrentthreadssoftlimit-gw2 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestchecktable-gw4 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestclusterallreplicas-gw1 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestallowedclienthosts-gw8 are DRIVER VOLUME NAME Docker containers for project roottestaltercodec-gw9 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Command:[docker container list --all --filter name='^/roottestallowedclienthosts-gw8-.*-1$' --format '{{.ID}}:{{.Names}}'] Docker volumes for project roottestconfigsubstitutions-gw3 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestconfigsubstitutions-gw3-.*-1$' --format '{{.ID}}:{{.Names}}'] Docker volumes for project roottestasyncmetricsincgroup-gw6 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestasyncmetricsincgroup-gw6-.*-1$' --format '{{.ID}}:{{.Names}}'] Docker volumes for project roottestbackuprestoreoncluster-gw0 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestbackuprestoreoncluster-gw0-.*-1$' --format '{{.ID}}:{{.Names}}'] Docker volumes for project roottestbackuprestore-gw5 are DRIVER VOLUME NAME Docker volumes for project roottestaccessforfunctions-gw7 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestbackuprestore-gw5-.*-1$' --format '{{.ID}}:{{.Names}}'] Command:[docker container list --all --filter name='^/roottestaccessforfunctions-gw7-.*-1$' --format '{{.ID}}:{{.Names}}'] Docker volumes for project roottestconcurrentthreadssoftlimit-gw2 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestconcurrentthreadssoftlimit-gw2-.*-1$' --format '{{.ID}}:{{.Names}}'] Docker volumes for project roottestchecktable-gw4 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestchecktable-gw4-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestallowedclienthosts-gw8 Trying to prune unused networks... Docker volumes for project roottestaltercodec-gw9 are DRIVER VOLUME NAME Docker volumes for project roottestclusterallreplicas-gw1 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestaltercodec-gw9-.*-1$' --format '{{.ID}}:{{.Names}}'] Command:[docker container list --all --filter name='^/roottestclusterallreplicas-gw1-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestconfigsubstitutions-gw3 Trying to prune unused networks... Unstopped containers: {} No running containers for project: roottestbackuprestoreoncluster-gw0 Trying to prune unused networks... Unstopped containers: {} No running containers for project: roottestconcurrentthreadssoftlimit-gw2 Trying to prune unused networks... Unstopped containers: {} No running containers for project: roottestasyncmetricsincgroup-gw6 Trying to prune unused networks... Unstopped containers: {} No running containers for project: roottestbackuprestore-gw5 Trying to prune unused networks... Unstopped containers: {} No running containers for project: roottestaccessforfunctions-gw7 Trying to prune unused networks... Unstopped containers: {} No running containers for project: roottestchecktable-gw4 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Unstopped containers: {} No running containers for project: roottestclusterallreplicas-gw1 Trying to prune unused networks... Unstopped containers: {} No running containers for project: roottestaltercodec-gw9 Trying to prune unused networks... Trying to prune unused images... Trying to prune unused images... Command:[docker image prune -f] Command:[docker image prune -f] Trying to prune unused images... Command:[docker image prune -f] Trying to prune unused images... Command:[docker image prune -f] Trying to prune unused images... Command:[docker image prune -f] Trying to prune unused images... Command:[docker image prune -f] Trying to prune unused images... Command:[docker image prune -f] Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stderr:Error response from daemon: a prune operation is already running Exitcode:1 Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stderr:Error response from daemon: a prune operation is already running Exitcode:1 Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stderr:Error response from daemon: a prune operation is already running Exitcode:1 Trying to prune unused volumes... Stdout:Total reclaimed space: 0B Images pruned Command:[docker volume ls | wc -l] Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stderr:Error response from daemon: a prune operation is already running Exitcode:1 Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:1 Stdout:1 Volumes pruned: 1 Setup directory for instance: node1 Volumes pruned: 1 Setup directory for instance: node1 Stdout:1 Volumes pruned: 1 Create directory for configuration generated in this helper Setup directory for instance: node1 Create directory for common tests configuration Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Copy common configuration from helpers Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Generate and write macros file Copy custom test config files [] to /ClickHouse/tests/integration/test_async_metrics_in_cgroup/_instances-0-gw6/node1/configs/config.d Generate and write macros file Setup database dir /ClickHouse/tests/integration/test_async_metrics_in_cgroup/_instances-0-gw6/node1/database Setup logs dir /ClickHouse/tests/integration/test_async_metrics_in_cgroup/_instances-0-gw6/node1/logs Entrypoint cmd: bash -c "trap 'pkill tail' INT TERM; clickhouse server --config-file=/etc/clickhouse-server/config.xml --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log --daemon -- ; coproc tail -f /dev/null; wait $$!" Copy custom test config files [] to /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node1/configs/config.d Copy custom test config files ['/ClickHouse/tests/integration/test_backup_restore_on_cluster/configs/cluster.xml', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/configs/cluster3.xml', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/configs/replicated_access_storage.xml', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/configs/replicated_user_defined_sql_objects.xml', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/configs/backups_disk.xml', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/configs/lesser_timeouts.xml'] to /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/node1/configs/config.d Setup directory for instance: node2 Setup database dir /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node1/database Create directory for configuration generated in this helper Create directory for common tests configuration Setup logs dir /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node1/logs Setup database dir /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/node1/database Copy common configuration from helpers Setup logs dir /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/node1/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Stdout:1 external_dir_abs_path=/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/backups Volumes pruned: 1 Generate and write macros file Setup directory for instance: node2 Setup directory for instance: server Create directory for configuration generated in this helper Copy custom test config files [] to /ClickHouse/tests/integration/test_async_metrics_in_cgroup/_instances-0-gw6/node2/configs/config.d Create directory for common tests configuration Setup database dir /ClickHouse/tests/integration/test_async_metrics_in_cgroup/_instances-0-gw6/node2/database Setup logs dir /ClickHouse/tests/integration/test_async_metrics_in_cgroup/_instances-0-gw6/node2/logs Copy common configuration from helpers Create directory for configuration generated in this helper Entrypoint cmd: bash -c "trap 'pkill tail' INT TERM; clickhouse server --config-file=/etc/clickhouse-server/config.xml --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log --daemon -- ; coproc tail -f /dev/null; wait $$!" Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw'} stored in /ClickHouse/tests/integration/test_async_metrics_in_cgroup/_instances-0-gw6/.env Create directory for common tests configuration Generate and write macros file Copy common configuration from helpers Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Copy custom test config files ['/ClickHouse/tests/integration/test_backup_restore_on_cluster/configs/cluster.xml', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/configs/cluster3.xml', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/configs/replicated_access_storage.xml', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/configs/replicated_user_defined_sql_objects.xml', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/configs/backups_disk.xml', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/configs/lesser_timeouts.xml'] to /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/node2/configs/config.d No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Generate and write macros file Setup database dir /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/node2/database Copy custom test config files [] to /ClickHouse/tests/integration/test_allowed_client_hosts/_instances-0-gw8/server/configs/config.d Setup logs dir /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/node2/logs Entrypoint cmd: bash -c "trap 'pkill tail' INT TERM; clickhouse server --config-file=/etc/clickhouse-server/config.xml --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log --daemon -- ; coproc tail -f /dev/null; wait $$!" external_dir_abs_path=/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/backups Setup database dir /ClickHouse/tests/integration/test_allowed_client_hosts/_instances-0-gw8/server/database Setup directory for instance: node3 Setup logs dir /ClickHouse/tests/integration/test_allowed_client_hosts/_instances-0-gw8/server/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Create directory for configuration generated in this helper Create directory for common tests configuration Setup directory for instance: clientA1 Copy common configuration from helpers Generate and write macros file Create directory for configuration generated in this helper Create directory for common tests configuration Copy custom test config files ['/ClickHouse/tests/integration/test_backup_restore_on_cluster/configs/cluster.xml', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/configs/cluster3.xml', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/configs/replicated_access_storage.xml', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/configs/replicated_user_defined_sql_objects.xml', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/configs/backups_disk.xml', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/configs/lesser_timeouts.xml'] to /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/node3/configs/config.d Copy common configuration from helpers Setup database dir /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/node3/database Setup logs dir /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/node3/logs Generate and write macros file Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] external_dir_abs_path=/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/backups Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'keeper_binary': '/clickhouse', 'keeper_cmd_prefix': 'clickhouse keeper', 'image': 'altinityinfra/integration-test:8b2301119731', 'user': '0', 'keeper_fs': 'bind', 'keeper_logs_dir1': '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/keeper1/log', 'keeper_config_dir1': '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/keeper1/config', 'keeper_db_dir1': '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/keeper1/coordination', 'keeper_logs_dir2': '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/keeper2/log', 'keeper_config_dir2': '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/keeper2/config', 'keeper_db_dir2': '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/keeper2/coordination', 'keeper_logs_dir3': '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/keeper3/log', 'keeper_config_dir3': '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/keeper3/config', 'keeper_db_dir3': '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/keeper3/coordination'} stored in /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/.env Copy custom test config files [] to /ClickHouse/tests/integration/test_allowed_client_hosts/_instances-0-gw8/clientA1/configs/config.d Setup database dir /ClickHouse/tests/integration/test_allowed_client_hosts/_instances-0-gw8/clientA1/database Setup logs dir /ClickHouse/tests/integration/test_allowed_client_hosts/_instances-0-gw8/clientA1/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Stdout:1 Setup directory for instance: clientA2 Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Volumes pruned: 1 No config file found Setup directory for instance: node1 Create directory for configuration generated in this helper Create directory for configuration generated in this helper Create directory for common tests configuration Create directory for common tests configuration Copy common configuration from helpers Copy common configuration from helpers Generate and write macros file Generate and write macros file Copy custom test config files [] to /ClickHouse/tests/integration/test_allowed_client_hosts/_instances-0-gw8/clientA2/configs/config.d Setup database dir /ClickHouse/tests/integration/test_allowed_client_hosts/_instances-0-gw8/clientA2/database Copy custom test config files ['/ClickHouse/tests/integration/test_concurrent_threads_soft_limit/configs/config_default.xml'] to /ClickHouse/tests/integration/test_concurrent_threads_soft_limit/_instances-0-gw2/node1/configs/config.d Setup logs dir /ClickHouse/tests/integration/test_allowed_client_hosts/_instances-0-gw8/clientA2/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Stdout:1 Setup database dir /ClickHouse/tests/integration/test_concurrent_threads_soft_limit/_instances-0-gw2/node1/database Setup logs dir /ClickHouse/tests/integration/test_concurrent_threads_soft_limit/_instances-0-gw2/node1/logs Setup directory for instance: clientA3 Volumes pruned: 1 Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Setup directory for instance: instance Setup directory for instance: node2 Create directory for configuration generated in this helper Create directory for configuration generated in this helper Create directory for configuration generated in this helper Create directory for common tests configuration Create directory for common tests configuration Create directory for common tests configuration Copy common configuration from helpers Copy common configuration from helpers Copy common configuration from helpers Generate and write macros file Copy custom test config files [] to /ClickHouse/tests/integration/test_access_for_functions/_instances-0-gw7/instance/configs/config.d Generate and write macros file Setup database dir /ClickHouse/tests/integration/test_access_for_functions/_instances-0-gw7/instance/database Generate and write macros file Setup logs dir /ClickHouse/tests/integration/test_access_for_functions/_instances-0-gw7/instance/logs Copy custom test config files [] to /ClickHouse/tests/integration/test_allowed_client_hosts/_instances-0-gw8/clientA3/configs/config.d Entrypoint cmd: bash -c "trap 'pkill tail' INT TERM; clickhouse server --config-file=/etc/clickhouse-server/config.xml --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log --daemon -- ; coproc tail -f /dev/null; wait $$!" Copy custom test config files ['/ClickHouse/tests/integration/test_concurrent_threads_soft_limit/configs/config_defined_50.xml'] to /ClickHouse/tests/integration/test_concurrent_threads_soft_limit/_instances-0-gw2/node2/configs/config.d Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw'} stored in /ClickHouse/tests/integration/test_access_for_functions/_instances-0-gw7/.env Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Setup database dir /ClickHouse/tests/integration/test_concurrent_threads_soft_limit/_instances-0-gw2/node2/database Setup logs dir /ClickHouse/tests/integration/test_concurrent_threads_soft_limit/_instances-0-gw2/node2/logs Setup directory for instance: node2 Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Setup directory for instance: node3 Stdout:1 Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Volumes pruned: 1 No config file found Setup directory for instance: node1 Create directory for configuration generated in this helper Create directory for configuration generated in this helper Create directory for common tests configuration Create directory for common tests configuration Copy common configuration from helpers Create directory for configuration generated in this helper Copy common configuration from helpers Create directory for common tests configuration Generate and write macros file http://localhost:None "GET /version HTTP/1.1" 200 826 Copy custom test config files ['/ClickHouse/tests/integration/test_concurrent_threads_soft_limit/configs/config_defined_1.xml'] to /ClickHouse/tests/integration/test_concurrent_threads_soft_limit/_instances-0-gw2/node3/configs/config.d Copy common configuration from helpers Setup database dir /ClickHouse/tests/integration/test_concurrent_threads_soft_limit/_instances-0-gw2/node3/database Generate and write macros file Setup logs dir /ClickHouse/tests/integration/test_concurrent_threads_soft_limit/_instances-0-gw2/node3/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Copy custom test config files [] to /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node2/configs/config.d Setup directory for instance: node4 Generate and write macros file Command:[docker compose --env-file /ClickHouse/tests/integration/test_async_metrics_in_cgroup/_instances-0-gw6/.env --project-name roottestasyncmetricsincgroup-gw6 --file /ClickHouse/tests/integration/test_async_metrics_in_cgroup/_instances-0-gw6/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_async_metrics_in_cgroup/_instances-0-gw6/node2/docker-compose.yml pull] Setup database dir /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node2/database Copy custom test config files ['/ClickHouse/tests/integration/test_cluster_all_replicas/configs/remote_servers.xml'] to /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw1/node1/configs/config.d Setup logs dir /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node2/logs Create directory for configuration generated in this helper Create directory for common tests configuration Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Setup database dir /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw1/node1/database Copy common configuration from helpers Setup directory for instance: node3 Setup logs dir /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw1/node1/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Stdout:1 Setup directory for instance: node2 Generate and write macros file Volumes pruned: 1 http://localhost:None "GET /version HTTP/1.1" 200 826 Create directory for configuration generated in this helper Copy custom test config files ['/ClickHouse/tests/integration/test_concurrent_threads_soft_limit/configs/config_limit_reached.xml'] to /ClickHouse/tests/integration/test_concurrent_threads_soft_limit/_instances-0-gw2/node4/configs/config.d Setup directory for instance: node1 Create directory for common tests configuration Create directory for configuration generated in this helper Setup database dir /ClickHouse/tests/integration/test_concurrent_threads_soft_limit/_instances-0-gw2/node4/database Create directory for common tests configuration Setup logs dir /ClickHouse/tests/integration/test_concurrent_threads_soft_limit/_instances-0-gw2/node4/logs Copy common configuration from helpers Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Copy common configuration from helpers Create directory for configuration generated in this helper Command:[docker compose --env-file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/.env --project-name roottestbackuprestoreoncluster-gw0 --file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/node2/docker-compose.yml --file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/node3/docker-compose.yml pull] Create directory for common tests configuration Setup database dir /ClickHouse/tests/integration/test_allowed_client_hosts/_instances-0-gw8/clientA3/database Copy common configuration from helpers Setup logs dir /ClickHouse/tests/integration/test_allowed_client_hosts/_instances-0-gw8/clientA3/logs Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw'} stored in /ClickHouse/tests/integration/test_concurrent_threads_soft_limit/_instances-0-gw2/.env Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Generate and write macros file Generate and write macros file Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw'} stored in /ClickHouse/tests/integration/test_allowed_client_hosts/_instances-0-gw8/.env Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_alter_codec/configs/logs_config.xml'] to /ClickHouse/tests/integration/test_alter_codec/_instances-0-gw9/node1/configs/config.d Copy custom test config files ['/ClickHouse/tests/integration/test_config_substitutions/configs/config_zk_include_test.xml'] to /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node3/configs/config.d Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Setup database dir /ClickHouse/tests/integration/test_alter_codec/_instances-0-gw9/node1/database No config file found Setup database dir /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node3/database Setup logs dir /ClickHouse/tests/integration/test_alter_codec/_instances-0-gw9/node1/logs Setup logs dir /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node3/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Copy custom test config files ['/ClickHouse/tests/integration/test_cluster_all_replicas/configs/remote_servers.xml'] to /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw1/node2/configs/config.d Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw'} stored in /ClickHouse/tests/integration/test_alter_codec/_instances-0-gw9/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Entrypoint cmd: bash -c "trap 'pkill tail' INT TERM; clickhouse server --config-file=/etc/clickhouse-server/config.xml --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log --daemon -- ; coproc tail -f /dev/null; wait $$!" No config file found Setup database dir /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw1/node2/database Setup directory for instance: node4 Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Setup logs dir /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw1/node2/logs No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] No config file found Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw'} stored in /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw1/.env Create directory for configuration generated in this helper Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Create directory for common tests configuration No config file found Copy common configuration from helpers Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_config_substitutions/configs/include_from_source.xml'] to /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node4/configs/config.d Setup database dir /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node4/database Setup logs dir /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node4/logs http://localhost:None "GET /version HTTP/1.1" 200 826 Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Setup directory for instance: node5 Create directory for configuration generated in this helper Create directory for common tests configuration Command:[docker compose --env-file /ClickHouse/tests/integration/test_access_for_functions/_instances-0-gw7/.env --project-name roottestaccessforfunctions-gw7 --file /ClickHouse/tests/integration/test_access_for_functions/_instances-0-gw7/instance/docker-compose.yml pull] Copy common configuration from helpers Generate and write macros file Copy custom test config files [] to /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node5/configs/config.d Setup database dir /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node5/database Setup logs dir /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node5/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Setup directory for instance: node6 Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_config_substitutions/configs/include_from_source.xml'] to /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node6/configs/config.d Setup database dir /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node6/database Setup logs dir /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node6/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Setup directory for instance: node7 Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_config_substitutions/configs/000-server_overrides.xml', '/ClickHouse/tests/integration/test_config_substitutions/configs/010-server_with_env_subst.xml'] to /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node7/configs/config.d Setup database dir /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node7/database Setup logs dir /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node7/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'MAX_QUERY_SIZE': '121212', 'keeper_binary': '/clickhouse', 'keeper_cmd_prefix': 'clickhouse keeper', 'image': 'altinityinfra/integration-test:8b2301119731', 'user': '0', 'keeper_fs': 'bind', 'keeper_logs_dir1': '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/keeper1/log', 'keeper_config_dir1': '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/keeper1/config', 'keeper_db_dir1': '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/keeper1/coordination', 'keeper_logs_dir2': '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/keeper2/log', 'keeper_config_dir2': '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/keeper2/config', 'keeper_db_dir2': '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/keeper2/coordination', 'keeper_logs_dir3': '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/keeper3/log', 'keeper_config_dir3': '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/keeper3/config', 'keeper_db_dir3': '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/keeper3/coordination', 'INCLUDE_FROM_ENV': '/etc/clickhouse-server/config.d/include_from_source.xml', 'MAX_THREADS': '2'} stored in /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node7/.env http://localhost:None "GET /version HTTP/1.1" 200 826 Setup directory for instance: node8 Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Command:[docker compose --env-file /ClickHouse/tests/integration/test_concurrent_threads_soft_limit/_instances-0-gw2/.env --project-name roottestconcurrentthreadssoftlimit-gw2 --file /ClickHouse/tests/integration/test_concurrent_threads_soft_limit/_instances-0-gw2/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_concurrent_threads_soft_limit/_instances-0-gw2/node2/docker-compose.yml --file /ClickHouse/tests/integration/test_concurrent_threads_soft_limit/_instances-0-gw2/node3/docker-compose.yml --file /ClickHouse/tests/integration/test_concurrent_threads_soft_limit/_instances-0-gw2/node4/docker-compose.yml pull] Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_config_substitutions/configs/include_from_source.yml'] to /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node8/configs/config.d Setup database dir /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node8/database Setup logs dir /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node8/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'MAX_QUERY_SIZE': '55555', 'keeper_binary': '/clickhouse', 'keeper_cmd_prefix': 'clickhouse keeper', 'image': 'altinityinfra/integration-test:8b2301119731', 'user': '0', 'keeper_fs': 'bind', 'keeper_logs_dir1': '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/keeper1/log', 'keeper_config_dir1': '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/keeper1/config', 'keeper_db_dir1': '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/keeper1/coordination', 'keeper_logs_dir2': '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/keeper2/log', 'keeper_config_dir2': '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/keeper2/config', 'keeper_db_dir2': '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/keeper2/coordination', 'keeper_logs_dir3': '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/keeper3/log', 'keeper_config_dir3': '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/keeper3/config', 'keeper_db_dir3': '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/keeper3/coordination', 'INCLUDE_FROM_ENV': '/etc/clickhouse-server/config.d/include_from_source.xml'} stored in /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Stdout:1 Volumes pruned: 1 Setup directory for instance: node Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files [] to /ClickHouse/tests/integration/test_backup_restore/_instances-0-gw5/node/configs/config.d Setup database dir /ClickHouse/tests/integration/test_backup_restore/_instances-0-gw5/node/database Setup logs dir /ClickHouse/tests/integration/test_backup_restore/_instances-0-gw5/node/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw'} stored in /ClickHouse/tests/integration/test_backup_restore/_instances-0-gw5/.env http://localhost:None "GET /version HTTP/1.1" 200 826 Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Command:[docker compose --env-file /ClickHouse/tests/integration/test_allowed_client_hosts/_instances-0-gw8/.env --project-name roottestallowedclienthosts-gw8 --file /ClickHouse/tests/integration/test_allowed_client_hosts/_instances-0-gw8/server/docker-compose.yml --file /ClickHouse/tests/integration/test_allowed_client_hosts/_instances-0-gw8/clientA1/docker-compose.yml --file /ClickHouse/tests/integration/test_allowed_client_hosts/_instances-0-gw8/clientA2/docker-compose.yml --file /ClickHouse/tests/integration/test_allowed_client_hosts/_instances-0-gw8/clientA3/docker-compose.yml pull] Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found http://localhost:None "GET /version HTTP/1.1" 200 826 Stdout:1 Volumes pruned: 1 http://localhost:None "GET /version HTTP/1.1" 200 826 Setup directory for instance: node1 Command:[docker compose --env-file /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw1/.env --project-name roottestclusterallreplicas-gw1 --file /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw1/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw1/node2/docker-compose.yml pull] Command:[docker compose --env-file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/.env --project-name roottestconfigsubstitutions-gw3 --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node2/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node3/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node4/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node5/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node6/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node7/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node8/docker-compose.yml pull] Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files [] to /ClickHouse/tests/integration/test_check_table/_instances-0-gw4/node1/configs/config.d Setup database dir /ClickHouse/tests/integration/test_check_table/_instances-0-gw4/node1/database Setup logs dir /ClickHouse/tests/integration/test_check_table/_instances-0-gw4/node1/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Setup directory for instance: node2 Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files [] to /ClickHouse/tests/integration/test_check_table/_instances-0-gw4/node2/configs/config.d Setup database dir /ClickHouse/tests/integration/test_check_table/_instances-0-gw4/node2/database Setup logs dir /ClickHouse/tests/integration/test_check_table/_instances-0-gw4/node2/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'keeper_binary': '/clickhouse', 'keeper_cmd_prefix': 'clickhouse keeper', 'image': 'altinityinfra/integration-test:8b2301119731', 'user': '0', 'keeper_fs': 'bind', 'keeper_logs_dir1': '/ClickHouse/tests/integration/test_check_table/_instances-0-gw4/keeper1/log', 'keeper_config_dir1': '/ClickHouse/tests/integration/test_check_table/_instances-0-gw4/keeper1/config', 'keeper_db_dir1': '/ClickHouse/tests/integration/test_check_table/_instances-0-gw4/keeper1/coordination', 'keeper_logs_dir2': '/ClickHouse/tests/integration/test_check_table/_instances-0-gw4/keeper2/log', 'keeper_config_dir2': '/ClickHouse/tests/integration/test_check_table/_instances-0-gw4/keeper2/config', 'keeper_db_dir2': '/ClickHouse/tests/integration/test_check_table/_instances-0-gw4/keeper2/coordination', 'keeper_logs_dir3': '/ClickHouse/tests/integration/test_check_table/_instances-0-gw4/keeper3/log', 'keeper_config_dir3': '/ClickHouse/tests/integration/test_check_table/_instances-0-gw4/keeper3/config', 'keeper_db_dir3': '/ClickHouse/tests/integration/test_check_table/_instances-0-gw4/keeper3/coordination'} stored in /ClickHouse/tests/integration/test_check_table/_instances-0-gw4/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_backup_restore/_instances-0-gw5/.env --project-name roottestbackuprestore-gw5 --file /ClickHouse/tests/integration/test_backup_restore/_instances-0-gw5/node/docker-compose.yml pull] http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_alter_codec/_instances-0-gw9/.env --project-name roottestaltercodec-gw9 --file /ClickHouse/tests/integration/test_alter_codec/_instances-0-gw9/node1/docker-compose.yml pull] http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_check_table/_instances-0-gw4/.env --project-name roottestchecktable-gw4 --file /ClickHouse/tests/integration/test_check_table/_instances-0-gw4/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_check_table/_instances-0-gw4/node2/docker-compose.yml pull] Stderr: zoo2 Skipped - Image is already being pulled by node3 Stderr: node1 Skipped - Image is already being pulled by node3 Stderr: node2 Skipped - Image is already being pulled by node3 Stderr: zoo1 Skipped - Image is already being pulled by node3 Stderr: node5 Skipped - Image is already being pulled by node3 Stderr: node8 Skipped - Image is already being pulled by node3 Stderr: node6 Skipped - Image is already being pulled by node3 Stderr: node4 Skipped - Image is already being pulled by node3 Stderr: zoo3 Skipped - Image is already being pulled by node3 Stderr: node7 Skipped - Image is already being pulled by node3 Stderr: node3 Pulling Stderr: node3 Pulled Setup ZooKeeper Creating internal ZooKeeper dirs: ['/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/keeper1/log', '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/keeper1/config', '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/keeper1/coordination', '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/keeper2/log', '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/keeper2/config', '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/keeper2/coordination', '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/keeper3/log', '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/keeper3/config', '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/keeper3/coordination'] Command:[docker compose --project-name roottestconfigsubstitutions-gw3 --env-file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --verbose up -d] Stderr: node1 Pulling Stderr: node1 Pulled ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_alter_codec/_instances-0-gw9/.env --project-name roottestaltercodec-gw9 --file /ClickHouse/tests/integration/test_alter_codec/_instances-0-gw9/node1/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_alter_codec/_instances-0-gw9/.env --project-name roottestaltercodec-gw9 --file /ClickHouse/tests/integration/test_alter_codec/_instances-0-gw9/node1/docker-compose.yml up -d --no-recreate] Stderr: node2 Skipped - Image is already being pulled by node1 Stderr: node3 Skipped - Image is already being pulled by node1 Stderr: node4 Skipped - Image is already being pulled by node1 Stderr: node1 Pulling Stderr: node1 Pulled ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_concurrent_threads_soft_limit/_instances-0-gw2/.env --project-name roottestconcurrentthreadssoftlimit-gw2 --file /ClickHouse/tests/integration/test_concurrent_threads_soft_limit/_instances-0-gw2/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_concurrent_threads_soft_limit/_instances-0-gw2/node2/docker-compose.yml --file /ClickHouse/tests/integration/test_concurrent_threads_soft_limit/_instances-0-gw2/node3/docker-compose.yml --file /ClickHouse/tests/integration/test_concurrent_threads_soft_limit/_instances-0-gw2/node4/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_concurrent_threads_soft_limit/_instances-0-gw2/.env --project-name roottestconcurrentthreadssoftlimit-gw2 --file /ClickHouse/tests/integration/test_concurrent_threads_soft_limit/_instances-0-gw2/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_concurrent_threads_soft_limit/_instances-0-gw2/node2/docker-compose.yml --file /ClickHouse/tests/integration/test_concurrent_threads_soft_limit/_instances-0-gw2/node3/docker-compose.yml --file /ClickHouse/tests/integration/test_concurrent_threads_soft_limit/_instances-0-gw2/node4/docker-compose.yml up -d --no-recreate] Stderr: zoo2 Skipped - Image is already being pulled by zoo1 Stderr: zoo3 Skipped - Image is already being pulled by zoo1 Stderr: node2 Skipped - Image is already being pulled by zoo1 Stderr: node1 Skipped - Image is already being pulled by zoo1 Stderr: zoo1 Pulling Stderr: zoo1 Pulled Setup ZooKeeper Creating internal ZooKeeper dirs: ['/ClickHouse/tests/integration/test_check_table/_instances-0-gw4/keeper1/log', '/ClickHouse/tests/integration/test_check_table/_instances-0-gw4/keeper1/config', '/ClickHouse/tests/integration/test_check_table/_instances-0-gw4/keeper1/coordination', '/ClickHouse/tests/integration/test_check_table/_instances-0-gw4/keeper2/log', '/ClickHouse/tests/integration/test_check_table/_instances-0-gw4/keeper2/config', '/ClickHouse/tests/integration/test_check_table/_instances-0-gw4/keeper2/coordination', '/ClickHouse/tests/integration/test_check_table/_instances-0-gw4/keeper3/log', '/ClickHouse/tests/integration/test_check_table/_instances-0-gw4/keeper3/config', '/ClickHouse/tests/integration/test_check_table/_instances-0-gw4/keeper3/coordination'] Command:[docker compose --project-name roottestchecktable-gw4 --env-file /ClickHouse/tests/integration/test_check_table/_instances-0-gw4/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --verbose up -d] Stderr: node1 Skipped - Image is already being pulled by node2 Stderr: node2 Pulling Stderr: node2 Pulled ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_async_metrics_in_cgroup/_instances-0-gw6/.env --project-name roottestasyncmetricsincgroup-gw6 --file /ClickHouse/tests/integration/test_async_metrics_in_cgroup/_instances-0-gw6/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_async_metrics_in_cgroup/_instances-0-gw6/node2/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_async_metrics_in_cgroup/_instances-0-gw6/.env --project-name roottestasyncmetricsincgroup-gw6 --file /ClickHouse/tests/integration/test_async_metrics_in_cgroup/_instances-0-gw6/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_async_metrics_in_cgroup/_instances-0-gw6/node2/docker-compose.yml up -d --no-recreate] Stderr: clientA3 Skipped - Image is already being pulled by clientA2 Stderr: server Skipped - Image is already being pulled by clientA2 Stderr: clientA1 Skipped - Image is already being pulled by clientA2 Stderr: clientA2 Pulling Stderr: clientA2 Pulled ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_allowed_client_hosts/_instances-0-gw8/.env --project-name roottestallowedclienthosts-gw8 --file /ClickHouse/tests/integration/test_allowed_client_hosts/_instances-0-gw8/server/docker-compose.yml --file /ClickHouse/tests/integration/test_allowed_client_hosts/_instances-0-gw8/clientA1/docker-compose.yml --file /ClickHouse/tests/integration/test_allowed_client_hosts/_instances-0-gw8/clientA2/docker-compose.yml --file /ClickHouse/tests/integration/test_allowed_client_hosts/_instances-0-gw8/clientA3/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_allowed_client_hosts/_instances-0-gw8/.env --project-name roottestallowedclienthosts-gw8 --file /ClickHouse/tests/integration/test_allowed_client_hosts/_instances-0-gw8/server/docker-compose.yml --file /ClickHouse/tests/integration/test_allowed_client_hosts/_instances-0-gw8/clientA1/docker-compose.yml --file /ClickHouse/tests/integration/test_allowed_client_hosts/_instances-0-gw8/clientA2/docker-compose.yml --file /ClickHouse/tests/integration/test_allowed_client_hosts/_instances-0-gw8/clientA3/docker-compose.yml up -d --no-recreate] Stderr: zoo1 Skipped - Image is already being pulled by node1 Stderr: zoo2 Skipped - Image is already being pulled by node1 Stderr: zoo3 Skipped - Image is already being pulled by node1 Stderr: node2 Skipped - Image is already being pulled by node1 Stderr: node3 Skipped - Image is already being pulled by node1 Stderr: node1 Pulling Stderr: node1 Pulled Setup ZooKeeper Creating internal ZooKeeper dirs: ['/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/keeper1/log', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/keeper1/config', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/keeper1/coordination', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/keeper2/log', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/keeper2/config', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/keeper2/coordination', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/keeper3/log', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/keeper3/config', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/keeper3/coordination'] Command:[docker compose --project-name roottestbackuprestoreoncluster-gw0 --env-file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --verbose up -d] Stderr: node Pulling Stderr: node Pulled ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_backup_restore/_instances-0-gw5/.env --project-name roottestbackuprestore-gw5 --file /ClickHouse/tests/integration/test_backup_restore/_instances-0-gw5/node/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_backup_restore/_instances-0-gw5/.env --project-name roottestbackuprestore-gw5 --file /ClickHouse/tests/integration/test_backup_restore/_instances-0-gw5/node/docker-compose.yml up -d --no-recreate] Stderr: node2 Skipped - Image is already being pulled by node1 Stderr: node1 Pulling Stderr: node1 Pulled ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw1/.env --project-name roottestclusterallreplicas-gw1 --file /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw1/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw1/node2/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw1/.env --project-name roottestclusterallreplicas-gw1 --file /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw1/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw1/node2/docker-compose.yml up -d --no-recreate] Stderr: instance Pulling Stderr: instance Pulled ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_access_for_functions/_instances-0-gw7/.env --project-name roottestaccessforfunctions-gw7 --file /ClickHouse/tests/integration/test_access_for_functions/_instances-0-gw7/instance/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_access_for_functions/_instances-0-gw7/.env --project-name roottestaccessforfunctions-gw7 --file /ClickHouse/tests/integration/test_access_for_functions/_instances-0-gw7/instance/docker-compose.yml up -d --no-recreate] Stderr: Network roottestaltercodec-gw9_default Creating Stderr: Network roottestaltercodec-gw9_default Created Stderr: Container roottestaltercodec-gw9-node1-1 Creating Stderr: Container roottestaltercodec-gw9-node1-1 Created Stderr: Container roottestaltercodec-gw9-node1-1 Starting Stderr: Container roottestaltercodec-gw9-node1-1 Started ClickHouse instance created get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottestaltercodec-gw9-node1-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottestaltercodec-gw9-node1-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node1, ip: 172.16.1.2... http://localhost:None "GET /v1.46/containers/roottestaltercodec-gw9-node1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b1f8a0cf3b9084e1b360b9257716cd2919015de13e2f80355a904155befdd5e8/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b1f8a0cf3b9084e1b360b9257716cd2919015de13e2f80355a904155befdd5e8/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b1f8a0cf3b9084e1b360b9257716cd2919015de13e2f80355a904155befdd5e8/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b1f8a0cf3b9084e1b360b9257716cd2919015de13e2f80355a904155befdd5e8/json HTTP/1.1" 200 None Stderr:time="2025-04-02T03:33:46Z" level=trace msg="Docker Desktop integration not enabled" Stderr: Network roottestconfigsubstitutions-gw3_default Creating Stderr: Network roottestconfigsubstitutions-gw3_default Created Stderr: Container roottestconfigsubstitutions-gw3-zoo3-1 Creating Stderr: Container roottestconfigsubstitutions-gw3-zoo1-1 Creating Stderr: Container roottestconfigsubstitutions-gw3-zoo2-1 Creating Stderr: Container roottestconfigsubstitutions-gw3-zoo1-1 Created Stderr: Container roottestconfigsubstitutions-gw3-zoo3-1 Created Stderr: Container roottestconfigsubstitutions-gw3-zoo2-1 Created Stderr: Container roottestconfigsubstitutions-gw3-zoo3-1 Starting Stderr: Container roottestconfigsubstitutions-gw3-zoo1-1 Starting Stderr: Container roottestconfigsubstitutions-gw3-zoo2-1 Starting Stderr: Container roottestconfigsubstitutions-gw3-zoo1-1 Started Stderr: Container roottestconfigsubstitutions-gw3-zoo3-1 Started Stderr: Container roottestconfigsubstitutions-gw3-zoo2-1 Started Stderr:time="2025-04-02T03:33:47Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T03:33:47Z" level=debug msg="otel error" error="" Wait ZooKeeper to start get_instance_ip instance_name=zoo1 http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw3-zoo1-1/json HTTP/1.1" 200 None get_kazoo_client: zoo1, ip:172.16.2.3, port:2181, use_ssl:False Connecting to 172.16.2.3(172.16.2.3):2181, use_ssl: False Connection dropped: socket connection error: Connection refused http://localhost:None "GET /v1.46/containers/b1f8a0cf3b9084e1b360b9257716cd2919015de13e2f80355a904155befdd5e8/json HTTP/1.1" 200 None ClickHouse node1 started Executing query CREATE TABLE test_alter_codec_index (`id` UInt64, value UInt64, INDEX id_index id TYPE minmax GRANULARITY 1) Engine=MergeTree() ORDER BY tuple() on node1 Connecting to 172.16.2.3(172.16.2.3):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Executing query INSERT INTO test_alter_codec_index SELECT number, number * number from numbers(100) on node1 Connecting to 172.16.2.3(172.16.2.3):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Executing query ALTER TABLE test_alter_codec_index MODIFY COLUMN id UInt64 CODEC(NONE) on node1 Connecting to 172.16.2.3(172.16.2.3):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Stderr:time="2025-04-02T03:33:46Z" level=trace msg="Docker Desktop integration not enabled" Stderr: Network roottestchecktable-gw4_default Creating Stderr: Network roottestchecktable-gw4_default Created Stderr: Container roottestchecktable-gw4-zoo2-1 Creating Stderr: Container roottestchecktable-gw4-zoo3-1 Creating Stderr: Container roottestchecktable-gw4-zoo1-1 Creating Stderr: Container roottestchecktable-gw4-zoo3-1 Created Stderr: Container roottestchecktable-gw4-zoo2-1 Created Stderr: Container roottestchecktable-gw4-zoo1-1 Created Stderr: Container roottestchecktable-gw4-zoo3-1 Starting Stderr: Container roottestchecktable-gw4-zoo2-1 Starting Stderr: Container roottestchecktable-gw4-zoo1-1 Starting Stderr: Container roottestchecktable-gw4-zoo3-1 Started Stderr: Container roottestchecktable-gw4-zoo1-1 Started Stderr: Container roottestchecktable-gw4-zoo2-1 Started Stderr:time="2025-04-02T03:33:47Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T03:33:47Z" level=debug msg="otel error" error="" Wait ZooKeeper to start get_instance_ip instance_name=zoo1 Executing query ALTER TABLE test_alter_codec_index MODIFY COLUMN id UInt64 CODEC(Delta, LZ4) on node1 http://localhost:None "GET /v1.46/containers/roottestchecktable-gw4-zoo1-1/json HTTP/1.1" 200 None get_kazoo_client: zoo1, ip:172.16.3.3, port:2181, use_ssl:False Connecting to 172.16.3.3(172.16.3.3):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Stderr: Network roottestconcurrentthreadssoftlimit-gw2_default Creating Stderr: Network roottestconcurrentthreadssoftlimit-gw2_default Created Stderr: Container roottestconcurrentthreadssoftlimit-gw2-node2-1 Creating Stderr: Container roottestconcurrentthreadssoftlimit-gw2-node3-1 Creating Stderr: Container roottestconcurrentthreadssoftlimit-gw2-node4-1 Creating Stderr: Container roottestconcurrentthreadssoftlimit-gw2-node1-1 Creating Stderr: Container roottestconcurrentthreadssoftlimit-gw2-node2-1 Created Stderr: Container roottestconcurrentthreadssoftlimit-gw2-node1-1 Created Stderr: Container roottestconcurrentthreadssoftlimit-gw2-node4-1 Created Stderr: Container roottestconcurrentthreadssoftlimit-gw2-node3-1 Created Stderr: Container roottestconcurrentthreadssoftlimit-gw2-node2-1 Starting Stderr: Container roottestconcurrentthreadssoftlimit-gw2-node3-1 Starting Stderr: Container roottestconcurrentthreadssoftlimit-gw2-node4-1 Starting Stderr: Container roottestconcurrentthreadssoftlimit-gw2-node1-1 Starting Stderr: Container roottestconcurrentthreadssoftlimit-gw2-node3-1 Started Stderr: Container roottestconcurrentthreadssoftlimit-gw2-node4-1 Started Stderr: Container roottestconcurrentthreadssoftlimit-gw2-node1-1 Started Stderr: Container roottestconcurrentthreadssoftlimit-gw2-node2-1 Started ClickHouse instance created get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottestconcurrentthreadssoftlimit-gw2-node1-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottestconcurrentthreadssoftlimit-gw2-node1-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node1, ip: 172.16.4.5... http://localhost:None "GET /v1.46/containers/roottestconcurrentthreadssoftlimit-gw2-node1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/4f751cd14ba3350142a3f91b1b6ce3c5392c063a1773ab9294ac4164f77c28f1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/4f751cd14ba3350142a3f91b1b6ce3c5392c063a1773ab9294ac4164f77c28f1/json HTTP/1.1" 200 None Connecting to 172.16.3.3(172.16.3.3):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Executing query ALTER TABLE test_alter_codec_index MODIFY COLUMN id UInt32 CODEC(Delta, LZ4) on node1 Executing query ALTER TABLE test_alter_codec_index MODIFY COLUMN id UInt64 DEFAULT 3 CODEC(Delta, LZ4) on node1 http://localhost:None "GET /v1.46/containers/4f751cd14ba3350142a3f91b1b6ce3c5392c063a1773ab9294ac4164f77c28f1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/4f751cd14ba3350142a3f91b1b6ce3c5392c063a1773ab9294ac4164f77c28f1/json HTTP/1.1" 200 None Executing query INSERT INTO test_alter_codec_index (value) VALUES (1) on node1 Stderr: Network roottestasyncmetricsincgroup-gw6_default Creating Stderr: Network roottestasyncmetricsincgroup-gw6_default Created Stderr: Container roottestasyncmetricsincgroup-gw6-node2-1 Creating Stderr: Container roottestasyncmetricsincgroup-gw6-node1-1 Creating Stderr: Container roottestasyncmetricsincgroup-gw6-node2-1 Created Stderr: Container roottestasyncmetricsincgroup-gw6-node1-1 Created Stderr: Container roottestasyncmetricsincgroup-gw6-node1-1 Starting Stderr: Container roottestasyncmetricsincgroup-gw6-node2-1 Starting Stderr: Container roottestasyncmetricsincgroup-gw6-node1-1 Started Stderr: Container roottestasyncmetricsincgroup-gw6-node2-1 Started ClickHouse instance created get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottestasyncmetricsincgroup-gw6-node1-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottestasyncmetricsincgroup-gw6-node1-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node1, ip: 172.16.5.2... http://localhost:None "GET /v1.46/containers/roottestasyncmetricsincgroup-gw6-node1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/1c679cebb3c6dd1cb2362afbb95b0a0b89a33c45cecff227828b867e86f601bc/json HTTP/1.1" 200 None Connecting to 172.16.2.3(172.16.2.3):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Connecting to 172.16.3.3(172.16.3.3):2181, use_ssl: False Connection dropped: socket connection error: Connection refused http://localhost:None "GET /v1.46/containers/4f751cd14ba3350142a3f91b1b6ce3c5392c063a1773ab9294ac4164f77c28f1/json HTTP/1.1" 200 None ClickHouse node1 started get_instance_ip instance_name=node2 http://localhost:None "GET /v1.46/containers/roottestconcurrentthreadssoftlimit-gw2-node2-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node2 http://localhost:None "GET /v1.46/containers/roottestconcurrentthreadssoftlimit-gw2-node2-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node2, ip: 172.16.4.3... http://localhost:None "GET /v1.46/containers/roottestconcurrentthreadssoftlimit-gw2-node2-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/f95bedbe9084f082c8a19522855848bc4175701ae989e99d77f50f0db9811508/json HTTP/1.1" 200 None ClickHouse node2 started get_instance_ip instance_name=node3 http://localhost:None "GET /v1.46/containers/roottestconcurrentthreadssoftlimit-gw2-node3-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node3 http://localhost:None "GET /v1.46/containers/roottestconcurrentthreadssoftlimit-gw2-node3-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node3, ip: 172.16.4.2... http://localhost:None "GET /v1.46/containers/roottestconcurrentthreadssoftlimit-gw2-node3-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/bb0fb4a21759d3903afab9eab349196cd2a39e508f4367c8e1945554d5756e75/json HTTP/1.1" 200 None ClickHouse node3 started get_instance_ip instance_name=node4 http://localhost:None "GET /v1.46/containers/roottestconcurrentthreadssoftlimit-gw2-node4-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node4 http://localhost:None "GET /v1.46/containers/roottestconcurrentthreadssoftlimit-gw2-node4-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node4, ip: 172.16.4.4... http://localhost:None "GET /v1.46/containers/roottestconcurrentthreadssoftlimit-gw2-node4-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/04792429457c764c6a393040b0c0a62bf5080518169780fe4341382d1835fa5a/json HTTP/1.1" 200 None ClickHouse node4 started Executing query SELECT count(*) FROM numbers_mt(10000000) on node1 http://localhost:None "GET /v1.46/containers/1c679cebb3c6dd1cb2362afbb95b0a0b89a33c45cecff227828b867e86f601bc/json HTTP/1.1" 200 None Stderr: Network roottestaccessforfunctions-gw7_default Creating Stderr: Network roottestaccessforfunctions-gw7_default Created Stderr: Container roottestaccessforfunctions-gw7-instance-1 Creating Stderr: Container roottestaccessforfunctions-gw7-instance-1 Created Stderr: Container roottestaccessforfunctions-gw7-instance-1 Starting Stderr: Container roottestaccessforfunctions-gw7-instance-1 Started ClickHouse instance created get_instance_ip instance_name=instance http://localhost:None "GET /v1.46/containers/roottestaccessforfunctions-gw7-instance-1/json HTTP/1.1" 200 None get_instance_ip instance_name=instance http://localhost:None "GET /v1.46/containers/roottestaccessforfunctions-gw7-instance-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in instance, ip: 172.16.7.2... http://localhost:None "GET /v1.46/containers/roottestaccessforfunctions-gw7-instance-1/json HTTP/1.1" 200 None Executing query SELECT sum(id) FROM test_alter_codec_index on node1 http://localhost:None "GET /v1.46/containers/307322e53e71c860f21d98aae4602a5409b811bb6f8dee7d3dd3115cb1449063/json HTTP/1.1" 200 None Stderr: Network roottestbackuprestore-gw5_default Creating Stderr: Network roottestbackuprestore-gw5_default Created Stderr: Container roottestbackuprestore-gw5-node-1 Creating Stderr: Container roottestbackuprestore-gw5-node-1 Created Stderr: Container roottestbackuprestore-gw5-node-1 Starting Stderr: Container roottestbackuprestore-gw5-node-1 Started ClickHouse instance created get_instance_ip instance_name=node http://localhost:None "GET /v1.46/containers/roottestbackuprestore-gw5-node-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node http://localhost:None "GET /v1.46/containers/roottestbackuprestore-gw5-node-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node, ip: 172.16.6.2... http://localhost:None "GET /v1.46/containers/roottestbackuprestore-gw5-node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b10a77903ec278f74644a1f7b3a328a27cf40441cdd7e9dcd421d55db89e6157/json HTTP/1.1" 200 None Executing query ALTER TABLE test_alter_codec_index MODIFY COLUMN id UInt64 ALIAS 3 CODEC(Delta, LZ4) on node1 http://localhost:None "GET /v1.46/containers/1c679cebb3c6dd1cb2362afbb95b0a0b89a33c45cecff227828b867e86f601bc/json HTTP/1.1" 200 None ClickHouse node1 started get_instance_ip instance_name=node2 http://localhost:None "GET /v1.46/containers/roottestasyncmetricsincgroup-gw6-node2-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node2 http://localhost:None "GET /v1.46/containers/roottestasyncmetricsincgroup-gw6-node2-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node2, ip: 172.16.5.3... http://localhost:None "GET /v1.46/containers/roottestasyncmetricsincgroup-gw6-node2-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/448d2d85e84a28e369cdf39d8033bf2abf990e6cf787d027207d289c8cc0bd56/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/307322e53e71c860f21d98aae4602a5409b811bb6f8dee7d3dd3115cb1449063/json HTTP/1.1" 200 None Executing query SYSTEM FLUSH LOGS on node1 http://localhost:None "GET /v1.46/containers/b10a77903ec278f74644a1f7b3a328a27cf40441cdd7e9dcd421d55db89e6157/json HTTP/1.1" 200 None Executing query ALTER TABLE test_alter_codec_index MODIFY COLUMN id UInt64 MATERIALIZED 3 CODEC(Delta, LZ4) on node1 http://localhost:None "GET /v1.46/containers/448d2d85e84a28e369cdf39d8033bf2abf990e6cf787d027207d289c8cc0bd56/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/307322e53e71c860f21d98aae4602a5409b811bb6f8dee7d3dd3115cb1449063/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b10a77903ec278f74644a1f7b3a328a27cf40441cdd7e9dcd421d55db89e6157/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/448d2d85e84a28e369cdf39d8033bf2abf990e6cf787d027207d289c8cc0bd56/json HTTP/1.1" 200 None Executing query INSERT INTO test_alter_codec_index (value) VALUES (1) on node1 Stderr: Network roottestclusterallreplicas-gw1_default Creating Stderr: Network roottestclusterallreplicas-gw1_default Created Stderr: Container roottestclusterallreplicas-gw1-node1-1 Creating Stderr: Container roottestclusterallreplicas-gw1-node2-1 Creating Stderr: Container roottestclusterallreplicas-gw1-node2-1 Created Stderr: Container roottestclusterallreplicas-gw1-node1-1 Created Stderr: Container roottestclusterallreplicas-gw1-node1-1 Starting Stderr: Container roottestclusterallreplicas-gw1-node2-1 Starting Stderr: Container roottestclusterallreplicas-gw1-node1-1 Started Stderr: Container roottestclusterallreplicas-gw1-node2-1 Started ClickHouse instance created get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottestclusterallreplicas-gw1-node1-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottestclusterallreplicas-gw1-node1-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node1, ip: 172.16.8.2... http://localhost:None "GET /v1.46/containers/roottestclusterallreplicas-gw1-node1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/307322e53e71c860f21d98aae4602a5409b811bb6f8dee7d3dd3115cb1449063/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b10a77903ec278f74644a1f7b3a328a27cf40441cdd7e9dcd421d55db89e6157/json HTTP/1.1" 200 None Executing query select ProfileEvents['ConcurrencyControlSlotsGranted'] from system.query_log where current_database = currentDatabase() and query_id = 'test_concurrent_threads_soft_limit_1' and type = 'QueryFinish' order by query_start_time_microseconds desc limit 1 on node1 http://localhost:None "GET /v1.46/containers/448d2d85e84a28e369cdf39d8033bf2abf990e6cf787d027207d289c8cc0bd56/json HTTP/1.1" 200 None ClickHouse node2 started Executing query SELECT value FROM system.build_options WHERE name = 'CXX_FLAGS' on node1 http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/307322e53e71c860f21d98aae4602a5409b811bb6f8dee7d3dd3115cb1449063/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b10a77903ec278f74644a1f7b3a328a27cf40441cdd7e9dcd421d55db89e6157/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/307322e53e71c860f21d98aae4602a5409b811bb6f8dee7d3dd3115cb1449063/json HTTP/1.1" 200 None ClickHouse instance started Executing query CREATE USER A on instance http://localhost:None "GET /v1.46/containers/b10a77903ec278f74644a1f7b3a328a27cf40441cdd7e9dcd421d55db89e6157/json HTTP/1.1" 200 None ClickHouse node started Executing query CREATE DATABASE test ENGINE = Ordinary on node Executing query select ProfileEvents['ConcurrencyControlSlotsDelayed'] from system.query_log where current_database = currentDatabase() and query_id = 'test_concurrent_threads_soft_limit_1' and type = 'QueryFinish' order by query_start_time_microseconds desc limit 1 on node1 Executing query SELECT sum(*) FROM system.numbers_mt FORMAT Null SETTINGS max_execution_time=10 on node2 Executing query SELECT sum(id) FROM test_alter_codec_index on node1 http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None Stderr: Network roottestallowedclienthosts-gw8_default Creating Stderr: Network roottestallowedclienthosts-gw8_default Created Stderr: Container roottestallowedclienthosts-gw8-clientA2-1 Creating Stderr: Container roottestallowedclienthosts-gw8-clientA3-1 Creating Stderr: Container roottestallowedclienthosts-gw8-clientA1-1 Creating Stderr: Container roottestallowedclienthosts-gw8-server-1 Creating Stderr: Container roottestallowedclienthosts-gw8-clientA3-1 Created Stderr: Container roottestallowedclienthosts-gw8-server-1 Created Stderr: Container roottestallowedclienthosts-gw8-clientA1-1 Created Stderr: Container roottestallowedclienthosts-gw8-clientA2-1 Created Stderr: Container roottestallowedclienthosts-gw8-clientA2-1 Starting Stderr: Container roottestallowedclienthosts-gw8-clientA3-1 Starting Stderr: Container roottestallowedclienthosts-gw8-server-1 Starting Stderr: Container roottestallowedclienthosts-gw8-clientA1-1 Starting Stderr: Container roottestallowedclienthosts-gw8-clientA2-1 Started Stderr: Container roottestallowedclienthosts-gw8-clientA1-1 Started Stderr: Container roottestallowedclienthosts-gw8-server-1 Started Stderr: Container roottestallowedclienthosts-gw8-clientA3-1 Started ClickHouse instance created get_instance_ip instance_name=server http://localhost:None "GET /v1.46/containers/roottestallowedclienthosts-gw8-server-1/json HTTP/1.1" 200 None get_instance_ip instance_name=server Connecting to 172.16.3.3(172.16.3.3):2181, use_ssl: False Connection dropped: socket connection error: Connection refused http://localhost:None "GET /v1.46/containers/roottestallowedclienthosts-gw8-server-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in server, ip: 172.16.9.3... http://localhost:None "GET /v1.46/containers/roottestallowedclienthosts-gw8-server-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/f1648366af9865e7fbdb961bad7f9a524412053b0ea8b4d674088158d1e88804/json HTTP/1.1" 200 None Executing query CREATE USER B on instance Executing query DROP TABLE IF EXISTS test.tbl on node Executing query select ProfileEvents['ConcurrencyControlSlotsAcquired'] from system.query_log where current_database = currentDatabase() and query_id = 'test_concurrent_threads_soft_limit_1' and type = 'QueryFinish' order by query_start_time_microseconds desc limit 1 on node1 Executing query ALTER TABLE test_alter_codec_index MODIFY COLUMN id UInt64 on node1 http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/f1648366af9865e7fbdb961bad7f9a524412053b0ea8b4d674088158d1e88804/json HTTP/1.1" 200 None Executing query ALTER TABLE test_alter_codec_index MODIFY COLUMN id Int64 on node1 Executing query CREATE TABLE test.tbl (p Date, k Int8) ENGINE = MergeTree PARTITION BY toYYYYMM(p) ORDER BY p on node Executing query select ProfileEvents['ConcurrencyControlSlotsAcquiredNonCompeting'] from system.query_log where current_database = currentDatabase() and query_id = 'test_concurrent_threads_soft_limit_1' and type = 'QueryFinish' order by query_start_time_microseconds desc limit 1 on node1 http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/f1648366af9865e7fbdb961bad7f9a524412053b0ea8b4d674088158d1e88804/json HTTP/1.1" 200 None Stderr:time="2025-04-02T03:33:46Z" level=trace msg="Docker Desktop integration not enabled" Stderr: Network roottestbackuprestoreoncluster-gw0_default Creating Stderr: Network roottestbackuprestoreoncluster-gw0_default Created Stderr: Container roottestbackuprestoreoncluster-gw0-zoo3-1 Creating Stderr: Container roottestbackuprestoreoncluster-gw0-zoo1-1 Creating Stderr: Container roottestbackuprestoreoncluster-gw0-zoo2-1 Creating Stderr: Container roottestbackuprestoreoncluster-gw0-zoo2-1 Created Stderr: Container roottestbackuprestoreoncluster-gw0-zoo1-1 Created Stderr: Container roottestbackuprestoreoncluster-gw0-zoo3-1 Created Stderr: Container roottestbackuprestoreoncluster-gw0-zoo1-1 Starting Stderr: Container roottestbackuprestoreoncluster-gw0-zoo2-1 Starting Stderr: Container roottestbackuprestoreoncluster-gw0-zoo3-1 Starting Stderr: Container roottestbackuprestoreoncluster-gw0-zoo1-1 Started Stderr: Container roottestbackuprestoreoncluster-gw0-zoo2-1 Started Stderr: Container roottestbackuprestoreoncluster-gw0-zoo3-1 Started Stderr:time="2025-04-02T03:33:49Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T03:33:49Z" level=debug msg="otel error" error="" Wait ZooKeeper to start get_instance_ip instance_name=zoo1 http://localhost:None "GET /v1.46/containers/roottestbackuprestoreoncluster-gw0-zoo1-1/json HTTP/1.1" 200 None get_kazoo_client: zoo1, ip:172.16.10.2, port:2181, use_ssl:False Connecting to 172.16.10.2(172.16.10.2):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Executing query CREATE FUNCTION MySum AS (a, b) -> a + b on instance Executing query select ProfileEvents['ConcurrencyControlQueriesDelayed'] from system.query_log where current_database = currentDatabase() and query_id = 'test_concurrent_threads_soft_limit_1' and type = 'QueryFinish' order by query_start_time_microseconds desc limit 1 on node1 Executing query DROP TABLE IF EXISTS test_alter_codec_index on node1 Executing query GRANT CREATE FUNCTION on *.* TO A on instance Executing query INSERT INTO test.tbl (p, k) VALUES(toDate(1), 1) on node Connecting to 172.16.2.3(172.16.2.3):2181, use_ssl: False Connection dropped: socket connection error: Connection refused http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None Executing query CREATE TABLE test_alter_codec_pk (id UInt64, value UInt64) Engine=MergeTree() ORDER BY id on node1 [gw9] PASSED test_alter_codec/test.py::test_alter_codec_index test_alter_codec/test.py::test_alter_codec_pk http://localhost:None "GET /v1.46/containers/f1648366af9865e7fbdb961bad7f9a524412053b0ea8b4d674088158d1e88804/json HTTP/1.1" 200 None Connecting to 172.16.10.2(172.16.10.2):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Executing query select length(thread_ids) from system.query_log where current_database = currentDatabase() and type = 'QueryFinish' and query_id = 'test_concurrent_threads_soft_limit_1' order by query_start_time_microseconds desc limit 1 on node1 Executing query CREATE FUNCTION MySum AS (a, b) -> a + b on instance Executing query INSERT INTO test.tbl (p, k) VALUES(toDate(2), 2) on node Executing query INSERT INTO test_alter_codec_pk SELECT number, number * number from numbers(100) on node1 Executing query SELECT MySum(1, 2) on instance Executing query SELECT count(*) FROM numbers_mt(10000000) on node3 [gw2] PASSED test_concurrent_threads_soft_limit/test.py::test_concurrent_threads_soft_limit_default test_concurrent_threads_soft_limit/test.py::test_concurrent_threads_soft_limit_defined_1 http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None Executing query INSERT INTO test.tbl (p, k) VALUES(toDate(3), 3) on node Executing query ALTER TABLE test_alter_codec_pk MODIFY COLUMN id UInt64 CODEC(NONE) on node1 http://localhost:None "GET /v1.46/containers/f1648366af9865e7fbdb961bad7f9a524412053b0ea8b4d674088158d1e88804/json HTTP/1.1" 200 None Executing query DROP FUNCTION MySum on instance Executing query SYSTEM FLUSH LOGS on node3 Executing query ALTER TABLE test_alter_codec_pk MODIFY COLUMN id UInt64 CODEC(Delta, LZ4) on node1 Executing query INSERT INTO test.tbl (p, k) VALUES(toDate(31), 31) on node http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None Executing query SELECT sum(id) FROM test_alter_codec_pk on node1 Executing query GRANT DROP FUNCTION ON *.* TO B on instance http://localhost:None "GET /v1.46/containers/f1648366af9865e7fbdb961bad7f9a524412053b0ea8b4d674088158d1e88804/json HTTP/1.1" 200 None ClickHouse server started get_instance_ip instance_name=clientA1 http://localhost:None "GET /v1.46/containers/roottestallowedclienthosts-gw8-clientA1-1/json HTTP/1.1" 200 None get_instance_ip instance_name=clientA1 http://localhost:None "GET /v1.46/containers/roottestallowedclienthosts-gw8-clientA1-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in clientA1, ip: 172.16.9.4... http://localhost:None "GET /v1.46/containers/roottestallowedclienthosts-gw8-clientA1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/40f5d7f76626261e74e83faa5439b4cd3230d32ed76b2aa5a2b93b22b8ddb717/json HTTP/1.1" 200 None ClickHouse clientA1 started get_instance_ip instance_name=clientA2 http://localhost:None "GET /v1.46/containers/roottestallowedclienthosts-gw8-clientA2-1/json HTTP/1.1" 200 None get_instance_ip instance_name=clientA2 http://localhost:None "GET /v1.46/containers/roottestallowedclienthosts-gw8-clientA2-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in clientA2, ip: 172.16.9.5... http://localhost:None "GET /v1.46/containers/roottestallowedclienthosts-gw8-clientA2-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/60ad8a4f984d83d6358e0eac78654d15e6ec2eb9067d12c67b3e0bd64dba4914/json HTTP/1.1" 200 None ClickHouse clientA2 started get_instance_ip instance_name=clientA3 http://localhost:None "GET /v1.46/containers/roottestallowedclienthosts-gw8-clientA3-1/json HTTP/1.1" 200 None get_instance_ip instance_name=clientA3 http://localhost:None "GET /v1.46/containers/roottestallowedclienthosts-gw8-clientA3-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in clientA3, ip: 172.16.9.2... http://localhost:None "GET /v1.46/containers/roottestallowedclienthosts-gw8-clientA3-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/ce1520c10b3a279b041fd006b5b54ea29654a45969fa8bab7a9d3022978b0e6c/json HTTP/1.1" 200 None Executing query DROP FUNCTION MySum on instance Executing query INSERT INTO test.tbl (p, k) VALUES(toDate(32), 32) on node Connecting to 172.16.10.2(172.16.10.2):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Executing query SELECT MySum(1, 2) on instance Executing query ALTER TABLE test_alter_codec_pk MODIFY COLUMN id UInt32 CODEC(Delta, LZ4) on node1 http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None Executing query INSERT INTO test.tbl (p, k) VALUES(toDate(33), 33) on node Executing query ALTER TABLE test_alter_codec_pk MODIFY COLUMN id UInt64 DEFAULT 3 CODEC(Delta, LZ4) on node1 http://localhost:None "GET /v1.46/containers/ce1520c10b3a279b041fd006b5b54ea29654a45969fa8bab7a9d3022978b0e6c/json HTTP/1.1" 200 None ClickHouse clientA3 started run container_id:roottestallowedclienthosts-gw8-server-1 detach:False nothrow:False cmd: ['bash', '-c', '/usr/bin/curl -s server:8123 '] Command:[docker exec roottestallowedclienthosts-gw8-server-1 bash -c /usr/bin/curl -s server:8123 ] Executing query REVOKE CREATE FUNCTION ON *.* FROM A on instance Executing query select ProfileEvents['ConcurrencyControlSlotsGranted'] from system.query_log where current_database = currentDatabase() and query_id = 'test_concurrent_threads_soft_limit_3' and type = 'QueryFinish' order by query_start_time_microseconds desc limit 1 on node3 Executing query SELECT * FROM test.tbl ORDER BY p on node Executing query INSERT INTO test_alter_codec_pk (value) VALUES (1) on node1 http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None Stdout:Ok. run container_id:roottestallowedclienthosts-gw8-server-1 detach:False nothrow:False cmd: ['ip', 'address', 'show'] Command:[docker exec roottestallowedclienthosts-gw8-server-1 ip address show] Executing query CREATE FUNCTION MySum AS (a, b) -> a + b on instance Executing query select ProfileEvents['ConcurrencyControlSlotsDelayed'] from system.query_log where current_database = currentDatabase() and query_id = 'test_concurrent_threads_soft_limit_3' and type = 'QueryFinish' order by query_start_time_microseconds desc limit 1 on node3 Stdout:1: lo: mtu 65536 qdisc noqueue state UNKNOWN group default qlen 1000 Stdout: link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 Stdout: inet 127.0.0.1/8 scope host lo Stdout: valid_lft forever preferred_lft forever Stdout: inet6 ::1/128 scope host Stdout: valid_lft forever preferred_lft forever Stdout:51: eth0@if52: mtu 1500 qdisc noqueue state UP group default Stdout: link/ether 02:42:ac:10:09:03 brd ff:ff:ff:ff:ff:ff link-netnsid 0 Stdout: inet 172.16.9.3/24 brd 172.16.9.255 scope global eth0 Stdout: valid_lft forever preferred_lft forever run container_id:roottestallowedclienthosts-gw8-server-1 detach:False nothrow:False cmd: ['host', 'clientA1.com'] Command:[docker exec roottestallowedclienthosts-gw8-server-1 host clientA1.com] Executing query SELECT sum(id) FROM test_alter_codec_pk on node1 Executing query DROP USER IF EXISTS A on instance Executing query ALTER TABLE test.tbl FREEZE on node http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None Executing query DROP USER IF EXISTS B on instance Connecting to 172.16.3.3(172.16.3.3):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost Executing query select ProfileEvents['ConcurrencyControlSlotsAcquired'] from system.query_log where current_database = currentDatabase() and query_id = 'test_concurrent_threads_soft_limit_3' and type = 'QueryFinish' order by query_start_time_microseconds desc limit 1 on node3 Stdout:clientA1.com has address 172.16.9.4 run container_id:roottestallowedclienthosts-gw8-server-1 detach:False nothrow:False cmd: ['host', 'clientA2.com'] Command:[docker exec roottestallowedclienthosts-gw8-server-1 host clientA2.com] Executing query CREATE TABLE test.tbl2 AS test.tbl on node Executing query ALTER TABLE test_alter_codec_pk MODIFY COLUMN id UInt64 ALIAS 3 CODEC(Delta, LZ4) on node1 run container_id:roottestaccessforfunctions-gw7-instance-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] Command:[docker exec -u root roottestaccessforfunctions-gw7-instance-1 bash -c ps -C clickhouse] [gw7] PASSED test_access_for_functions/test.py::test_access_rights_for_function test_access_for_functions/test.py::test_ignore_obsolete_grant_on_database http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None Executing query select ProfileEvents['ConcurrencyControlSlotsAcquiredNonCompeting'] from system.query_log where current_database = currentDatabase() and query_id = 'test_concurrent_threads_soft_limit_3' and type = 'QueryFinish' order by query_start_time_microseconds desc limit 1 on node3 Stdout:clientA2.com has address 172.16.9.5 run container_id:roottestallowedclienthosts-gw8-server-1 detach:False nothrow:False cmd: ['host', 'clientA3.com'] Command:[docker exec roottestallowedclienthosts-gw8-server-1 host clientA3.com] Executing query ALTER TABLE test_alter_codec_pk MODIFY COLUMN id UInt64 MATERIALIZED 3 CODEC(Delta, LZ4) on node1 Executing query INSERT INTO test.tbl2(p, k) VALUES(toDate(3), 3) on node Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED get_instance_ip instance_name=zoo2 http://localhost:None "GET /v1.46/containers/roottestchecktable-gw4-zoo2-1/json HTTP/1.1" 200 None get_kazoo_client: zoo2, ip:172.16.3.4, port:2181, use_ssl:False Stdout: PID TTY TIME CMD Stdout: 10 ? 00:00:00 clickhouse run container_id:roottestaccessforfunctions-gw7-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] Command:[docker exec -u root roottestaccessforfunctions-gw7-instance-1 bash -c pkill clickhouse] Connecting to 172.16.3.4(172.16.3.4):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost Executing query select ProfileEvents['ConcurrencyControlQueriesDelayed'] from system.query_log where current_database = currentDatabase() and query_id = 'test_concurrent_threads_soft_limit_3' and type = 'QueryFinish' order by query_start_time_microseconds desc limit 1 on node3 run container_id:roottestaccessforfunctions-gw7-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestaccessforfunctions-gw7-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query INSERT INTO test_alter_codec_pk (value) VALUES (1) on node1 Executing query INSERT INTO test.tbl2(p, k) VALUES(toDate(4), 4) on node http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None Stdout:clientA3.com has address 172.16.9.2 IP: 1: lo: mtu 65536 qdisc noqueue state UNKNOWN group default qlen 1000 link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 inet 127.0.0.1/8 scope host lo valid_lft forever preferred_lft forever inet6 ::1/128 scope host valid_lft forever preferred_lft forever 51: eth0@if52: mtu 1500 qdisc noqueue state UP group default link/ether 02:42:ac:10:09:03 brd ff:ff:ff:ff:ff:ff link-netnsid 0 inet 172.16.9.3/24 brd 172.16.9.255 scope global eth0 valid_lft forever preferred_lft forever , A1 clientA1.com has address 172.16.9.4 , A2 clientA2.com has address 172.16.9.5 , A3 clientA3.com has address 172.16.9.2 run container_id:roottestallowedclienthosts-gw8-server-1 detach:False nothrow:False cmd: ['bash', '-c', "/usr/bin/clickhouse client --host server --query 'DROP TABLE IF EXISTS test_allowed_client_hosts'"] Command:[docker exec roottestallowedclienthosts-gw8-server-1 bash -c /usr/bin/clickhouse client --host server --query 'DROP TABLE IF EXISTS test_allowed_client_hosts'] Connecting to 172.16.10.2(172.16.10.2):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Stdout:10 Executing query select length(thread_ids) from system.query_log where current_database = currentDatabase() and type = 'QueryFinish' and query_id = 'test_concurrent_threads_soft_limit_3' order by query_start_time_microseconds desc limit 1 on node3 Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED get_instance_ip instance_name=zoo3 http://localhost:None "GET /v1.46/containers/roottestchecktable-gw4-zoo3-1/json HTTP/1.1" 200 None get_kazoo_client: zoo3, ip:172.16.3.2, port:2181, use_ssl:False Connecting to 172.16.3.2(172.16.3.2):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Executing query SELECT sum(id) FROM test_alter_codec_pk on node1 Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Executing query INSERT INTO test.tbl2(p, k) VALUES(toDate(33), 33) on node Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None run container_id:roottestallowedclienthosts-gw8-server-1 detach:False nothrow:False cmd: ['bash', '-c', '/usr/bin/curl -s server:8123 '] Command:[docker exec roottestallowedclienthosts-gw8-server-1 bash -c /usr/bin/curl -s server:8123 ] Executing query SELECT count(*) FROM numbers_mt(10000000) on node2 [gw2] PASSED test_concurrent_threads_soft_limit/test.py::test_concurrent_threads_soft_limit_defined_1 test_concurrent_threads_soft_limit/test.py::test_concurrent_threads_soft_limit_defined_50 Executing query INSERT INTO test.tbl2(p, k) VALUES(toDate(34), 34) on node Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED All instances of ZooKeeper started: ('zoo1', 'zoo2', 'zoo3') ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_check_table/_instances-0-gw4/.env --project-name roottestchecktable-gw4 --file /ClickHouse/tests/integration/test_check_table/_instances-0-gw4/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_check_table/_instances-0-gw4/node2/docker-compose.yml up -d --no-recreate') Stdout:Ok. Command:[docker compose --env-file /ClickHouse/tests/integration/test_check_table/_instances-0-gw4/.env --project-name roottestchecktable-gw4 --file /ClickHouse/tests/integration/test_check_table/_instances-0-gw4/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_check_table/_instances-0-gw4/node2/docker-compose.yml up -d --no-recreate] run container_id:roottestallowedclienthosts-gw8-server-1 detach:False nothrow:False cmd: ['ip', 'address', 'show'] Command:[docker exec roottestallowedclienthosts-gw8-server-1 ip address show] Executing query ALTER TABLE test_alter_codec_pk MODIFY COLUMN id UInt64 on node1 Executing query SELECT * FROM test.tbl2 ORDER BY p on node http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None Executing query ALTER TABLE test_alter_codec_pk MODIFY COLUMN id Int64 on node1 Executing query SYSTEM FLUSH LOGS on node2 Stdout:1: lo: mtu 65536 qdisc noqueue state UNKNOWN group default qlen 1000 Stdout: link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 Stdout: inet 127.0.0.1/8 scope host lo Stdout: valid_lft forever preferred_lft forever Stdout: inet6 ::1/128 scope host Stdout: valid_lft forever preferred_lft forever Stdout:51: eth0@if52: mtu 1500 qdisc noqueue state UP group default Stdout: link/ether 02:42:ac:10:09:03 brd ff:ff:ff:ff:ff:ff link-netnsid 0 Stdout: inet 172.16.9.3/24 brd 172.16.9.255 scope global eth0 Stdout: valid_lft forever preferred_lft forever run container_id:roottestallowedclienthosts-gw8-server-1 detach:False nothrow:False cmd: ['host', 'clientA1.com'] Command:[docker exec roottestallowedclienthosts-gw8-server-1 host clientA1.com] Executing query DROP TABLE IF EXISTS test_alter_codec_pk on node1 http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None copy from /var/lib/clickhouse/shadow/*/data/test/tbl to /var/lib/clickhouse/data/test/tbl2/detached run container_id:roottestbackuprestore-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', 'cp -r /var/lib/clickhouse/shadow/*/data/test/tbl -T /var/lib/clickhouse/data/test/tbl2/detached'] Command:[docker exec roottestbackuprestore-gw5-node-1 bash -c cp -r /var/lib/clickhouse/shadow/*/data/test/tbl -T /var/lib/clickhouse/data/test/tbl2/detached] Stdout:clientA1.com has address 172.16.9.4 run container_id:roottestallowedclienthosts-gw8-server-1 detach:False nothrow:False cmd: ['host', 'clientA2.com'] Command:[docker exec roottestallowedclienthosts-gw8-server-1 host clientA2.com] Command:[docker compose --env-file /ClickHouse/tests/integration/test_alter_codec/_instances-0-gw9/.env --project-name roottestaltercodec-gw9 --file /ClickHouse/tests/integration/test_alter_codec/_instances-0-gw9/node1/docker-compose.yml stop --timeout 20] Executing query ALTER TABLE test.tbl2 ATTACH PARTITION 197001 on node [gw9] PASSED test_alter_codec/test.py::test_alter_codec_pk http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None Executing query ALTER TABLE test.tbl2 ATTACH PARTITION 197002 on node Executing query select ProfileEvents['ConcurrencyControlSlotsGranted'] from system.query_log where current_database = currentDatabase() and query_id = 'test_concurrent_threads_soft_limit_2' and type = 'QueryFinish' order by query_start_time_microseconds desc limit 1 on node2 Stdout:clientA2.com has address 172.16.9.5 run container_id:roottestallowedclienthosts-gw8-server-1 detach:False nothrow:False cmd: ['host', 'clientA3.com'] Command:[docker exec roottestallowedclienthosts-gw8-server-1 host clientA3.com] Executing query SELECT sleep(2) on node Executing query select ProfileEvents['ConcurrencyControlSlotsDelayed'] from system.query_log where current_database = currentDatabase() and query_id = 'test_concurrent_threads_soft_limit_2' and type = 'QueryFinish' order by query_start_time_microseconds desc limit 1 on node2 Stdout:clientA3.com has address 172.16.9.2 IP: 1: lo: mtu 65536 qdisc noqueue state UNKNOWN group default qlen 1000 link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 inet 127.0.0.1/8 scope host lo valid_lft forever preferred_lft forever inet6 ::1/128 scope host valid_lft forever preferred_lft forever 51: eth0@if52: mtu 1500 qdisc noqueue state UP group default link/ether 02:42:ac:10:09:03 brd ff:ff:ff:ff:ff:ff link-netnsid 0 inet 172.16.9.3/24 brd 172.16.9.255 scope global eth0 valid_lft forever preferred_lft forever , A1 clientA1.com has address 172.16.9.4 , A2 clientA2.com has address 172.16.9.5 , A3 clientA3.com has address 172.16.9.2 run container_id:roottestallowedclienthosts-gw8-server-1 detach:False nothrow:False cmd: ['bash', '-c', "/usr/bin/clickhouse client --host server --query 'CREATE TABLE test_allowed_client_hosts (x Int32) ENGINE = MergeTree() ORDER BY tuple()'"] Command:[docker exec roottestallowedclienthosts-gw8-server-1 bash -c /usr/bin/clickhouse client --host server --query 'CREATE TABLE test_allowed_client_hosts (x Int32) ENGINE = MergeTree() ORDER BY tuple()'] http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None Executing query select ProfileEvents['ConcurrencyControlSlotsAcquired'] from system.query_log where current_database = currentDatabase() and query_id = 'test_concurrent_threads_soft_limit_2' and type = 'QueryFinish' order by query_start_time_microseconds desc limit 1 on node2 http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None run container_id:roottestallowedclienthosts-gw8-server-1 detach:False nothrow:False cmd: ['bash', '-c', '/usr/bin/curl -s server:8123 '] Command:[docker exec roottestallowedclienthosts-gw8-server-1 bash -c /usr/bin/curl -s server:8123 ] Executing query select ProfileEvents['ConcurrencyControlSlotsAcquiredNonCompeting'] from system.query_log where current_database = currentDatabase() and query_id = 'test_concurrent_threads_soft_limit_2' and type = 'QueryFinish' order by query_start_time_microseconds desc limit 1 on node2 http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None Stdout:Ok. run container_id:roottestallowedclienthosts-gw8-server-1 detach:False nothrow:False cmd: ['ip', 'address', 'show'] Command:[docker exec roottestallowedclienthosts-gw8-server-1 ip address show] Executing query select ProfileEvents['ConcurrencyControlQueriesDelayed'] from system.query_log where current_database = currentDatabase() and query_id = 'test_concurrent_threads_soft_limit_2' and type = 'QueryFinish' order by query_start_time_microseconds desc limit 1 on node2 Stderr: Container roottestchecktable-gw4-zoo2-1 Running Stderr: Container roottestchecktable-gw4-zoo3-1 Running Stderr: Container roottestchecktable-gw4-zoo1-1 Running Stderr: Container roottestchecktable-gw4-node1-1 Creating Stderr: Container roottestchecktable-gw4-node2-1 Creating Stderr: Container roottestchecktable-gw4-node1-1 Created Stderr: Container roottestchecktable-gw4-node2-1 Created Stderr: Container roottestchecktable-gw4-node1-1 Starting Stderr: Container roottestchecktable-gw4-node2-1 Starting Stderr: Container roottestchecktable-gw4-node1-1 Started Stderr: Container roottestchecktable-gw4-node2-1 Started ClickHouse instance created get_instance_ip instance_name=node1 Stdout:1: lo: mtu 65536 qdisc noqueue state UNKNOWN group default qlen 1000 Stdout: link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 Stdout: inet 127.0.0.1/8 scope host lo Stdout: valid_lft forever preferred_lft forever Stdout: inet6 ::1/128 scope host Stdout: valid_lft forever preferred_lft forever Stdout:51: eth0@if52: mtu 1500 qdisc noqueue state UP group default Stdout: link/ether 02:42:ac:10:09:03 brd ff:ff:ff:ff:ff:ff link-netnsid 0 Stdout: inet 172.16.9.3/24 brd 172.16.9.255 scope global eth0 Stdout: valid_lft forever preferred_lft forever run container_id:roottestallowedclienthosts-gw8-server-1 detach:False nothrow:False cmd: ['host', 'clientA1.com'] Command:[docker exec roottestallowedclienthosts-gw8-server-1 host clientA1.com] http://localhost:None "GET /v1.46/containers/roottestchecktable-gw4-node1-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottestchecktable-gw4-node1-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node1, ip: 172.16.3.5... http://localhost:None "GET /v1.46/containers/roottestchecktable-gw4-node1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/9d7ec31f2fd06f4bd791e836419599d3a64f41983cc71b487afe3ee93ace7dad/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/9d7ec31f2fd06f4bd791e836419599d3a64f41983cc71b487afe3ee93ace7dad/json HTTP/1.1" 200 None Executing query select length(thread_ids) from system.query_log where current_database = currentDatabase() and type = 'QueryFinish' and query_id = 'test_concurrent_threads_soft_limit_2' order by query_start_time_microseconds desc limit 1 on node2 http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None Stdout:clientA1.com has address 172.16.9.4 run container_id:roottestallowedclienthosts-gw8-server-1 detach:False nothrow:False cmd: ['host', 'clientA2.com'] Command:[docker exec roottestallowedclienthosts-gw8-server-1 host clientA2.com] Executing query SELECT count(*) FROM numbers_mt(1e11) settings max_threads=100 on node4 [gw2] PASSED test_concurrent_threads_soft_limit/test.py::test_concurrent_threads_soft_limit_defined_50 test_concurrent_threads_soft_limit/test.py::test_concurrent_threads_soft_limit_limit_reached Executing query SELECT sum(length(thread_ids)) FROM system.processes on node4 http://localhost:None "GET /v1.46/containers/9d7ec31f2fd06f4bd791e836419599d3a64f41983cc71b487afe3ee93ace7dad/json HTTP/1.1" 200 None Stdout:clientA2.com has address 172.16.9.5 run container_id:roottestallowedclienthosts-gw8-server-1 detach:False nothrow:False cmd: ['host', 'clientA3.com'] Command:[docker exec roottestallowedclienthosts-gw8-server-1 host clientA3.com] http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None run container_id:roottestaccessforfunctions-gw7-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestaccessforfunctions-gw7-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Connecting to 172.16.10.2(172.16.10.2):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Stdout:10 Stdout:clientA3.com has address 172.16.9.2 IP: 1: lo: mtu 65536 qdisc noqueue state UNKNOWN group default qlen 1000 link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 inet 127.0.0.1/8 scope host lo valid_lft forever preferred_lft forever inet6 ::1/128 scope host valid_lft forever preferred_lft forever 51: eth0@if52: mtu 1500 qdisc noqueue state UP group default link/ether 02:42:ac:10:09:03 brd ff:ff:ff:ff:ff:ff link-netnsid 0 inet 172.16.9.3/24 brd 172.16.9.255 scope global eth0 valid_lft forever preferred_lft forever , A1 clientA1.com has address 172.16.9.4 , A2 clientA2.com has address 172.16.9.5 , A3 clientA3.com has address 172.16.9.2 run container_id:roottestallowedclienthosts-gw8-server-1 detach:False nothrow:False cmd: ['bash', '-c', "/usr/bin/clickhouse client --host server --query 'INSERT INTO test_allowed_client_hosts VALUES (5)'"] Command:[docker exec roottestallowedclienthosts-gw8-server-1 bash -c /usr/bin/clickhouse client --host server --query 'INSERT INTO test_allowed_client_hosts VALUES (5)'] http://localhost:None "GET /v1.46/containers/9d7ec31f2fd06f4bd791e836419599d3a64f41983cc71b487afe3ee93ace7dad/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None Executing query SELECT count(*) FROM numbers_mt(10000000) settings max_threads=6 on node4 run container_id:roottestallowedclienthosts-gw8-server-1 detach:False nothrow:False cmd: ['bash', '-c', '/usr/bin/curl -s server:8123 '] Command:[docker exec roottestallowedclienthosts-gw8-server-1 bash -c /usr/bin/curl -s server:8123 ] http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/9d7ec31f2fd06f4bd791e836419599d3a64f41983cc71b487afe3ee93ace7dad/json HTTP/1.1" 200 None Stdout:Ok. run container_id:roottestallowedclienthosts-gw8-server-1 detach:False nothrow:False cmd: ['ip', 'address', 'show'] Command:[docker exec roottestallowedclienthosts-gw8-server-1 ip address show] http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None Stdout:1: lo: mtu 65536 qdisc noqueue state UNKNOWN group default qlen 1000 Executing query SYSTEM FLUSH LOGS on node4 Stdout: link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 Stdout: inet 127.0.0.1/8 scope host lo Stdout: valid_lft forever preferred_lft forever Stdout: inet6 ::1/128 scope host Stdout: valid_lft forever preferred_lft forever Stdout:51: eth0@if52: mtu 1500 qdisc noqueue state UP group default Stdout: link/ether 02:42:ac:10:09:03 brd ff:ff:ff:ff:ff:ff link-netnsid 0 Stdout: inet 172.16.9.3/24 brd 172.16.9.255 scope global eth0 Stdout: valid_lft forever preferred_lft forever run container_id:roottestallowedclienthosts-gw8-server-1 detach:False nothrow:False cmd: ['host', 'clientA1.com'] Command:[docker exec roottestallowedclienthosts-gw8-server-1 host clientA1.com] http://localhost:None "GET /v1.46/containers/9d7ec31f2fd06f4bd791e836419599d3a64f41983cc71b487afe3ee93ace7dad/json HTTP/1.1" 200 None Connecting to 172.16.2.3(172.16.2.3):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost Stdout:clientA1.com has address 172.16.9.4 run container_id:roottestallowedclienthosts-gw8-server-1 detach:False nothrow:False cmd: ['host', 'clientA2.com'] Command:[docker exec roottestallowedclienthosts-gw8-server-1 host clientA2.com] http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/9d7ec31f2fd06f4bd791e836419599d3a64f41983cc71b487afe3ee93ace7dad/json HTTP/1.1" 200 None ClickHouse node1 started get_instance_ip instance_name=node2 http://localhost:None "GET /v1.46/containers/roottestchecktable-gw4-node2-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node2 http://localhost:None "GET /v1.46/containers/roottestchecktable-gw4-node2-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node2, ip: 172.16.3.6... http://localhost:None "GET /v1.46/containers/roottestchecktable-gw4-node2-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/734322af55c32f92cb569bf0676e28583eae0f75c36f765d1c6abdc8cabcc84b/json HTTP/1.1" 200 None ClickHouse node2 started Executing query CREATE DATABASE db1 on node1 Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED get_instance_ip instance_name=zoo2 http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw3-zoo2-1/json HTTP/1.1" 200 None get_kazoo_client: zoo2, ip:172.16.2.4, port:2181, use_ssl:False Connecting to 172.16.2.4(172.16.2.4):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Stdout:clientA2.com has address 172.16.9.5 run container_id:roottestallowedclienthosts-gw8-server-1 detach:False nothrow:False cmd: ['host', 'clientA3.com'] Command:[docker exec roottestallowedclienthosts-gw8-server-1 host clientA3.com] Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None Executing query CREATE TABLE db1.table1 (date Date, id UInt32, value Int32) ENGINE = MergeTree() PARTITION BY toYYYYMM(date) ORDER BY id SETTINGS min_bytes_for_wide_part=0 on node1 Stdout:clientA3.com has address 172.16.9.2 IP: 1: lo: mtu 65536 qdisc noqueue state UNKNOWN group default qlen 1000 link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 inet 127.0.0.1/8 scope host lo valid_lft forever preferred_lft forever inet6 ::1/128 scope host valid_lft forever preferred_lft forever 51: eth0@if52: mtu 1500 qdisc noqueue state UP group default link/ether 02:42:ac:10:09:03 brd ff:ff:ff:ff:ff:ff link-netnsid 0 inet 172.16.9.3/24 brd 172.16.9.255 scope global eth0 valid_lft forever preferred_lft forever , A1 clientA1.com has address 172.16.9.4 , A2 clientA2.com has address 172.16.9.5 , A3 clientA3.com has address 172.16.9.2 run container_id:roottestallowedclienthosts-gw8-server-1 detach:False nothrow:False cmd: ['bash', '-c', "/usr/bin/clickhouse client --host server --query 'SELECT fqdn(), hostName()'"] Command:[docker exec roottestallowedclienthosts-gw8-server-1 bash -c /usr/bin/clickhouse client --host server --query 'SELECT fqdn(), hostName()'] Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED get_instance_ip instance_name=zoo3 http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw3-zoo3-1/json HTTP/1.1" 200 None get_kazoo_client: zoo3, ip:172.16.2.2, port:2181, use_ssl:False Connecting to 172.16.2.2(172.16.2.2):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Executing query select ProfileEvents['ConcurrencyControlSlotsGranted'] from system.query_log where current_database = currentDatabase() and query_id = 'test_concurrent_threads_soft_limit_4' and type = 'QueryFinish' order by query_start_time_microseconds desc limit 1 on node4 Sending request(xid=2): Close() Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None Executing query INSERT INTO db1.table1 VALUES (toDate('2019-02-01'), 1, 10), (toDate('2019-02-01'), 2, 12) on node1 Executing query select ProfileEvents['ConcurrencyControlSlotsDelayed'] from system.query_log where current_database = currentDatabase() and query_id = 'test_concurrent_threads_soft_limit_4' and type = 'QueryFinish' order by query_start_time_microseconds desc limit 1 on node4 Stdout:server server run container_id:roottestallowedclienthosts-gw8-clientA1-1 detach:False nothrow:False cmd: ['bash', '-c', '/usr/bin/curl -s clientA1.com:8123 '] Command:[docker exec roottestallowedclienthosts-gw8-clientA1-1 bash -c /usr/bin/curl -s clientA1.com:8123 ] Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED All instances of ZooKeeper started: ('zoo1', 'zoo2', 'zoo3') get_instance_ip instance_name=zoo1 Executing query SYSTEM STOP MERGES db1.table1 on node1 http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw3-zoo1-1/json HTTP/1.1" 200 None get_kazoo_client: zoo1, ip:172.16.2.3, port:2181, use_ssl:False Connecting to 172.16.2.3(172.16.2.3):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED run_kazoo_commands_with_retries: zoo1, .create_zk_roots at 0x7fe1c1c5a170> Sending request(xid=1): Create(path='/setting/max_query_size', data=b'77777', acl=[ACL(perms=31, acl_list=['ALL'], id=Id(scheme='world', id='anyone'))], flags=0) Received error(xid=1) NoNodeError() Sending request(xid=2): Exists(path='/setting', watcher=None) Sending request(xid=3): Exists(path='/', watcher=None) Received response(xid=3): ZnodeStat(czxid=0, mzxid=0, ctime=0, mtime=0, version=0, cversion=0, aversion=0, ephemeralOwner=0, dataLength=0, numChildren=1, pzxid=0) Sending request(xid=4): Create(path='/setting', data=b'', acl=[ACL(perms=31, acl_list=['ALL'], id=Id(scheme='world', id='anyone'))], flags=0) Received response(xid=4): '/setting' Sending request(xid=5): Create(path='/setting/max_query_size', data=b'77777', acl=[ACL(perms=31, acl_list=['ALL'], id=Id(scheme='world', id='anyone'))], flags=0) Received response(xid=5): '/setting/max_query_size' Sending request(xid=6): Create(path='/users_from_zk_1', data=b'default', acl=[ACL(perms=31, acl_list=['ALL'], id=Id(scheme='world', id='anyone'))], flags=0) Received response(xid=6): '/users_from_zk_1' Sending request(xid=7): Create(path='/users_from_zk_2', data=b'default', acl=[ACL(perms=31, acl_list=['ALL'], id=Id(scheme='world', id='anyone'))], flags=0) Received response(xid=7): '/users_from_zk_2' http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None Sending request(xid=8): Create(path='/min_bytes_for_wide_part', data=b'33', acl=[ACL(perms=31, acl_list=['ALL'], id=Id(scheme='world', id='anyone'))], flags=0) Received response(xid=8): '/min_bytes_for_wide_part' Sending request(xid=9): Create(path='/merge_max_block_size', data=b'8888', acl=[ACL(perms=31, acl_list=['ALL'], id=Id(scheme='world', id='anyone'))], flags=0) Received response(xid=9): '/merge_max_block_size' ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/.env --project-name roottestconfigsubstitutions-gw3 --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node2/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node3/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node4/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node5/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node6/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node7/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node8/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/.env --project-name roottestconfigsubstitutions-gw3 --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node2/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node3/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node4/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node5/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node6/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node7/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node8/docker-compose.yml up -d --no-recreate] Executing query select ProfileEvents['ConcurrencyControlSlotsAcquired'] from system.query_log where current_database = currentDatabase() and query_id = 'test_concurrent_threads_soft_limit_4' and type = 'QueryFinish' order by query_start_time_microseconds desc limit 1 on node4 Stdout:Ok. run container_id:roottestallowedclienthosts-gw8-clientA1-1 detach:False nothrow:False cmd: ['ip', 'address', 'show'] Command:[docker exec roottestallowedclienthosts-gw8-clientA1-1 ip address show] Executing query CREATE TABLE db1.table2 (date Date, id UInt32, value Int32) ENGINE = MergeTree() PARTITION BY toYYYYMM(date) ORDER BY id SETTINGS min_bytes_for_wide_part=0 on node1 http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None Executing query INSERT INTO db1.table2 VALUES (toDate('2019-02-01'), 1, 10), (toDate('2019-02-01'), 2, 12) on node1 Stdout:1: lo: mtu 65536 qdisc noqueue state UNKNOWN group default qlen 1000 Stdout: link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 Stdout: inet 127.0.0.1/8 scope host lo Stdout: valid_lft forever preferred_lft forever Stdout: inet6 ::1/128 scope host Stdout: valid_lft forever preferred_lft forever Stdout:53: eth0@if54: mtu 1500 qdisc noqueue state UP group default Stdout: link/ether 02:42:ac:10:09:04 brd ff:ff:ff:ff:ff:ff link-netnsid 0 Stdout: inet 172.16.9.4/24 brd 172.16.9.255 scope global eth0 Stdout: valid_lft forever preferred_lft forever run container_id:roottestallowedclienthosts-gw8-clientA1-1 detach:False nothrow:False cmd: ['host', 'clientA1.com'] Command:[docker exec roottestallowedclienthosts-gw8-clientA1-1 host clientA1.com] Executing query select ProfileEvents['ConcurrencyControlSlotsAcquiredNonCompeting'] from system.query_log where current_database = currentDatabase() and query_id = 'test_concurrent_threads_soft_limit_4' and type = 'QueryFinish' order by query_start_time_microseconds desc limit 1 on node4 Executing query SYSTEM STOP MERGES db1.table2 on node1 http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None Executing query CREATE DATABASE db2 on node1 Executing query select ProfileEvents['ConcurrencyControlQueriesDelayed'] from system.query_log where current_database = currentDatabase() and query_id = 'test_concurrent_threads_soft_limit_4' and type = 'QueryFinish' order by query_start_time_microseconds desc limit 1 on node4 Stdout:clientA1.com has address 172.16.9.4 run container_id:roottestallowedclienthosts-gw8-clientA1-1 detach:False nothrow:False cmd: ['host', 'clientA2.com'] Command:[docker exec roottestallowedclienthosts-gw8-clientA1-1 host clientA2.com] Executing query CREATE TABLE db2.table1 (date Date, id UInt32, value Int32) ENGINE = MergeTree() PARTITION BY toYYYYMM(date) ORDER BY id SETTINGS min_bytes_for_wide_part=0 on node1 http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None Executing query select length(thread_ids) from system.query_log where current_database = currentDatabase() and type = 'QueryFinish' and query_id = 'test_concurrent_threads_soft_limit_4' order by query_start_time_microseconds desc limit 1 on node4 run container_id:roottestaccessforfunctions-gw7-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestaccessforfunctions-gw7-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:clientA2.com has address 172.16.9.5 run container_id:roottestallowedclienthosts-gw8-clientA1-1 detach:False nothrow:False cmd: ['host', 'clientA3.com'] Command:[docker exec roottestallowedclienthosts-gw8-clientA1-1 host clientA3.com] Executing query INSERT INTO db2.table1 VALUES (toDate('2019-02-01'), 1, 10), (toDate('2019-02-01'), 2, 12) on node1 http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None Executing query KILL QUERY WHERE query_id = 'background_query' SYNC on node4 Stdout:10 Executing query SYSTEM STOP MERGES db2.table1 on node1 Stdout:clientA3.com has address 172.16.9.2 IP: 1: lo: mtu 65536 qdisc noqueue state UNKNOWN group default qlen 1000 link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 inet 127.0.0.1/8 scope host lo valid_lft forever preferred_lft forever inet6 ::1/128 scope host valid_lft forever preferred_lft forever 53: eth0@if54: mtu 1500 qdisc noqueue state UP group default link/ether 02:42:ac:10:09:04 brd ff:ff:ff:ff:ff:ff link-netnsid 0 inet 172.16.9.4/24 brd 172.16.9.255 scope global eth0 valid_lft forever preferred_lft forever , A1 clientA1.com has address 172.16.9.4 , A2 clientA2.com has address 172.16.9.5 , A3 clientA3.com has address 172.16.9.2 run container_id:roottestallowedclienthosts-gw8-clientA1-1 detach:False nothrow:False cmd: ['bash', '-c', "/usr/bin/clickhouse client --host clientA1.com --query 'SELECT fqdn(), hostName()'"] Command:[docker exec roottestallowedclienthosts-gw8-clientA1-1 bash -c /usr/bin/clickhouse client --host clientA1.com --query 'SELECT fqdn(), hostName()'] http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None Executing query CREATE TABLE db2.table2 (date Date, id UInt32, value Int32) ENGINE = MergeTree() PARTITION BY toYYYYMM(date) ORDER BY id SETTINGS min_bytes_for_wide_part=0 on node1 Executing query SELECT count(*) FROM numbers_mt(10000000) SETTINGS use_concurrency_control = 0 on node1 [gw2] PASSED test_concurrent_threads_soft_limit/test.py::test_concurrent_threads_soft_limit_limit_reached test_concurrent_threads_soft_limit/test.py::test_use_concurrency_control_default Executing query INSERT INTO db2.table2 VALUES (toDate('2019-02-01'), 1, 10), (toDate('2019-02-01'), 2, 12) on node1 http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None Stdout:clientA1.com clientA1.com run container_id:roottestallowedclienthosts-gw8-clientA2-1 detach:False nothrow:False cmd: ['bash', '-c', '/usr/bin/curl -s clientA2.com:8123 '] Command:[docker exec roottestallowedclienthosts-gw8-clientA2-1 bash -c /usr/bin/curl -s clientA2.com:8123 ] Stdout:Ok. run container_id:roottestallowedclienthosts-gw8-clientA2-1 detach:False nothrow:False cmd: ['ip', 'address', 'show'] Command:[docker exec roottestallowedclienthosts-gw8-clientA2-1 ip address show] http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None Executing query SYSTEM STOP MERGES db2.table2 on node1 Executing query SELECT * FROM test.tbl2 ORDER BY p on node Executing query SYSTEM FLUSH LOGS on node1 Stdout:1: lo: mtu 65536 qdisc noqueue state UNKNOWN group default qlen 1000 Stdout: link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 Stdout: inet 127.0.0.1/8 scope host lo Stdout: valid_lft forever preferred_lft forever Stdout: inet6 ::1/128 scope host Stdout: valid_lft forever preferred_lft forever Stdout:55: eth0@if56: mtu 1500 qdisc noqueue state UP group default Stdout: link/ether 02:42:ac:10:09:05 brd ff:ff:ff:ff:ff:ff link-netnsid 0 Stdout: inet 172.16.9.5/24 brd 172.16.9.255 scope global eth0 Stdout: valid_lft forever preferred_lft forever http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None run container_id:roottestallowedclienthosts-gw8-clientA2-1 detach:False nothrow:False cmd: ['host', 'clientA1.com'] Command:[docker exec roottestallowedclienthosts-gw8-clientA2-1 host clientA1.com] Stdout:clientA1.com has address 172.16.9.4 run container_id:roottestallowedclienthosts-gw8-clientA2-1 detach:False nothrow:False cmd: ['host', 'clientA2.com'] Command:[docker exec roottestallowedclienthosts-gw8-clientA2-1 host clientA2.com] Connecting to 172.16.10.2(172.16.10.2):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None Stdout:clientA2.com has address 172.16.9.5 run container_id:roottestallowedclienthosts-gw8-clientA2-1 detach:False nothrow:False cmd: ['host', 'clientA3.com'] Command:[docker exec roottestallowedclienthosts-gw8-clientA2-1 host clientA3.com] Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED get_instance_ip instance_name=zoo2 http://localhost:None "GET /v1.46/containers/roottestbackuprestoreoncluster-gw0-zoo2-1/json HTTP/1.1" 200 None get_kazoo_client: zoo2, ip:172.16.10.3, port:2181, use_ssl:False Connecting to 172.16.10.3(172.16.10.3):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost Stdout:clientA3.com has address 172.16.9.2 IP: 1: lo: mtu 65536 qdisc noqueue state UNKNOWN group default qlen 1000 link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 inet 127.0.0.1/8 scope host lo valid_lft forever preferred_lft forever inet6 ::1/128 scope host valid_lft forever preferred_lft forever 55: eth0@if56: mtu 1500 qdisc noqueue state UP group default link/ether 02:42:ac:10:09:05 brd ff:ff:ff:ff:ff:ff link-netnsid 0 inet 172.16.9.5/24 brd 172.16.9.255 scope global eth0 valid_lft forever preferred_lft forever , A1 clientA1.com has address 172.16.9.4 , A2 clientA2.com has address 172.16.9.5 , A3 clientA3.com has address 172.16.9.2 run container_id:roottestallowedclienthosts-gw8-clientA2-1 detach:False nothrow:False cmd: ['bash', '-c', "/usr/bin/clickhouse client --host clientA2.com --query 'SELECT fqdn(), hostName()'"] Command:[docker exec roottestallowedclienthosts-gw8-clientA2-1 bash -c /usr/bin/clickhouse client --host clientA2.com --query 'SELECT fqdn(), hostName()'] http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED get_instance_ip instance_name=zoo3 http://localhost:None "GET /v1.46/containers/roottestbackuprestoreoncluster-gw0-zoo3-1/json HTTP/1.1" 200 None get_kazoo_client: zoo3, ip:172.16.10.4, port:2181, use_ssl:False Connecting to 172.16.10.4(172.16.10.4):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Executing query CREATE DATABASE db3 on node1 Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost Executing query ALTER TABLE test.tbl2 UPDATE k=10 WHERE 1 on node http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None Executing query select ProfileEvents['ConcurrencyControlSlotsGranted'] from system.query_log where current_database = currentDatabase() and query_id = 'test_use_concurrency_control' and type = 'QueryFinish' order by query_start_time_microseconds desc limit 1 on node1 Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED All instances of ZooKeeper started: ('zoo1', 'zoo2', 'zoo3') ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/.env --project-name roottestbackuprestoreoncluster-gw0 --file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/node2/docker-compose.yml --file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/node3/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/.env --project-name roottestbackuprestoreoncluster-gw0 --file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/node2/docker-compose.yml --file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/node3/docker-compose.yml up -d --no-recreate] Executing query select ProfileEvents['ConcurrencyControlSlotsDelayed'] from system.query_log where current_database = currentDatabase() and query_id = 'test_use_concurrency_control' and type = 'QueryFinish' order by query_start_time_microseconds desc limit 1 on node1 Stdout:clientA2.com clientA2.com run container_id:roottestallowedclienthosts-gw8-clientA3-1 detach:False nothrow:False cmd: ['bash', '-c', '/usr/bin/curl -s clientA3.com:8123 '] Command:[docker exec roottestallowedclienthosts-gw8-clientA3-1 bash -c /usr/bin/curl -s clientA3.com:8123 ] Executing query CREATE TABLE db3.table1 (date Date, id UInt32, value Int32) ENGINE = MergeTree() PARTITION BY toYYYYMM(date) ORDER BY id SETTINGS min_bytes_for_wide_part=0 on node1 http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None Executing query SELECT sleep(2) on node Stderr: Container roottestconfigsubstitutions-gw3-zoo1-1 Running Stderr: Container roottestconfigsubstitutions-gw3-zoo3-1 Running Stderr: Container roottestconfigsubstitutions-gw3-node6-1 Creating Stderr: Container roottestconfigsubstitutions-gw3-node8-1 Creating Stderr: Container roottestconfigsubstitutions-gw3-node2-1 Creating Stderr: Container roottestconfigsubstitutions-gw3-node7-1 Creating Stderr: Container roottestconfigsubstitutions-gw3-node1-1 Creating Stderr: Container roottestconfigsubstitutions-gw3-node5-1 Creating Stderr: Container roottestconfigsubstitutions-gw3-zoo2-1 Running Stderr: Container roottestconfigsubstitutions-gw3-node3-1 Creating Stderr: Container roottestconfigsubstitutions-gw3-node4-1 Creating Stderr: Container roottestconfigsubstitutions-gw3-node7-1 Created Stderr: Container roottestconfigsubstitutions-gw3-node5-1 Created Stderr: Container roottestconfigsubstitutions-gw3-node2-1 Created Stderr: Container roottestconfigsubstitutions-gw3-node8-1 Created Stderr: Container roottestconfigsubstitutions-gw3-node4-1 Created Stderr: Container roottestconfigsubstitutions-gw3-node3-1 Created Stderr: Container roottestconfigsubstitutions-gw3-node1-1 Created Stderr: Container roottestconfigsubstitutions-gw3-node6-1 Created Stderr: Container roottestconfigsubstitutions-gw3-node7-1 Starting Stderr: Container roottestconfigsubstitutions-gw3-node8-1 Starting Stderr: Container roottestconfigsubstitutions-gw3-node4-1 Starting Stderr: Container roottestconfigsubstitutions-gw3-node2-1 Starting Stderr: Container roottestconfigsubstitutions-gw3-node3-1 Starting Stderr: Container roottestconfigsubstitutions-gw3-node6-1 Starting Stderr: Container roottestconfigsubstitutions-gw3-node1-1 Starting Stderr: Container roottestconfigsubstitutions-gw3-node5-1 Starting Stderr: Container roottestconfigsubstitutions-gw3-node6-1 Started Stderr: Container roottestconfigsubstitutions-gw3-node8-1 Started Stderr: Container roottestconfigsubstitutions-gw3-node4-1 Started Stderr: Container roottestconfigsubstitutions-gw3-node2-1 Started Stderr: Container roottestconfigsubstitutions-gw3-node7-1 Started Stderr: Container roottestconfigsubstitutions-gw3-node3-1 Started Stderr: Container roottestconfigsubstitutions-gw3-node1-1 Started Stderr: Container roottestconfigsubstitutions-gw3-node5-1 Started ClickHouse instance created get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw3-node1-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw3-node1-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node1, ip: 172.16.2.12... http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw3-node1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/f4288bcc543c09f5b91784408f69373fa57bbff9651125c7cf2d841bebe0085a/json HTTP/1.1" 200 None Executing query INSERT INTO db3.table1 VALUES (toDate('2019-02-01'), 1, 10), (toDate('2019-02-01'), 2, 12) on node1 Stdout:Ok. run container_id:roottestallowedclienthosts-gw8-clientA3-1 detach:False nothrow:False cmd: ['ip', 'address', 'show'] Command:[docker exec roottestallowedclienthosts-gw8-clientA3-1 ip address show] http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None run container_id:roottestaccessforfunctions-gw7-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestaccessforfunctions-gw7-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query SYSTEM STOP MERGES db3.table1 on node1 Executing query select ProfileEvents['ConcurrencyControlSlotsAcquired'] from system.query_log where current_database = currentDatabase() and query_id = 'test_use_concurrency_control' and type = 'QueryFinish' order by query_start_time_microseconds desc limit 1 on node1 http://localhost:None "GET /v1.46/containers/f4288bcc543c09f5b91784408f69373fa57bbff9651125c7cf2d841bebe0085a/json HTTP/1.1" 200 None Stdout:1: lo: mtu 65536 qdisc noqueue state UNKNOWN group default qlen 1000 Stdout: link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 Stdout: inet 127.0.0.1/8 scope host lo Stdout: valid_lft forever preferred_lft forever Stdout: inet6 ::1/128 scope host Stdout: valid_lft forever preferred_lft forever Stdout:48: eth0@if49: mtu 1500 qdisc noqueue state UP group default Stdout: link/ether 02:42:ac:10:09:02 brd ff:ff:ff:ff:ff:ff link-netnsid 0 Stdout: inet 172.16.9.2/24 brd 172.16.9.255 scope global eth0 Stdout: valid_lft forever preferred_lft forever run container_id:roottestallowedclienthosts-gw8-clientA3-1 detach:False nothrow:False cmd: ['host', 'clientA1.com'] Command:[docker exec roottestallowedclienthosts-gw8-clientA3-1 host clientA1.com] Executing query CREATE TABLE db3.table2 (date Date, id UInt32, value Int32) ENGINE = MergeTree() PARTITION BY toYYYYMM(date) ORDER BY id SETTINGS min_bytes_for_wide_part=0 on node1 http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None Executing query select ProfileEvents['ConcurrencyControlSlotsAcquiredNonCompeting'] from system.query_log where current_database = currentDatabase() and query_id = 'test_use_concurrency_control' and type = 'QueryFinish' order by query_start_time_microseconds desc limit 1 on node1 run container_id:roottestaccessforfunctions-gw7-instance-1 detach:False nothrow:False cmd: ['bash', '-c', '\n cat > /var/lib/clickhouse/access/d481e934-6745-42eb-86e7-1d8aec307361.sql << EOF\nATTACH USER \\`d481e934-6745-42eb-86e7-1d8aec307361\\`;\nATTACH GRANT CREATE FUNCTION, SELECT ON mydb.* TO \\`d481e934-6745-42eb-86e7-1d8aec307361\\`;\nEOF'] Command:[docker exec roottestaccessforfunctions-gw7-instance-1 bash -c cat > /var/lib/clickhouse/access/d481e934-6745-42eb-86e7-1d8aec307361.sql << EOF ATTACH USER \`d481e934-6745-42eb-86e7-1d8aec307361\`; ATTACH GRANT CREATE FUNCTION, SELECT ON mydb.* TO \`d481e934-6745-42eb-86e7-1d8aec307361\`; EOF] http://localhost:None "GET /v1.46/containers/f4288bcc543c09f5b91784408f69373fa57bbff9651125c7cf2d841bebe0085a/json HTTP/1.1" 200 None Executing query INSERT INTO db3.table2 VALUES (toDate('2019-02-01'), 1, 10), (toDate('2019-02-01'), 2, 12) on node1 http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None Executing query select ProfileEvents['ConcurrencyControlQueriesDelayed'] from system.query_log where current_database = currentDatabase() and query_id = 'test_use_concurrency_control' and type = 'QueryFinish' order by query_start_time_microseconds desc limit 1 on node1 Stdout:clientA1.com has address 172.16.9.4 run container_id:roottestallowedclienthosts-gw8-clientA3-1 detach:False nothrow:False cmd: ['host', 'clientA2.com'] Command:[docker exec roottestallowedclienthosts-gw8-clientA3-1 host clientA2.com] run container_id:roottestaccessforfunctions-gw7-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'touch /var/lib/clickhouse/access/need_rebuild_lists.mark'] Command:[docker exec roottestaccessforfunctions-gw7-instance-1 bash -c touch /var/lib/clickhouse/access/need_rebuild_lists.mark] Executing query SYSTEM STOP MERGES db3.table2 on node1 http://localhost:None "GET /v1.46/containers/f4288bcc543c09f5b91784408f69373fa57bbff9651125c7cf2d841bebe0085a/json HTTP/1.1" 200 None run container_id:roottestaccessforfunctions-gw7-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestaccessforfunctions-gw7-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:clientA2.com has address 172.16.9.5 run container_id:roottestallowedclienthosts-gw8-clientA3-1 detach:False nothrow:False cmd: ['host', 'clientA3.com'] Executing query SELECT path FROM system.parts WHERE database = 'db1' AND table = 'table2' AND name = '201902_1_1_0' on node1 Command:[docker exec roottestallowedclienthosts-gw8-clientA3-1 host clientA3.com] http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None No clickhouse process running. Start new one. http://localhost:None "POST /v1.46/containers/roottestaccessforfunctions-gw7-instance-1/exec HTTP/1.1" 201 74 http://localhost:None "POST /v1.46/exec/b1da4b3f71e4fd669030c9154325c5961e50c4ca39892815d9cf3f742a464d8c/start HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/exec/b1da4b3f71e4fd669030c9154325c5961e50c4ca39892815d9cf3f742a464d8c/json HTTP/1.1" 200 586 http://localhost:None "GET /v1.46/containers/f4288bcc543c09f5b91784408f69373fa57bbff9651125c7cf2d841bebe0085a/json HTTP/1.1" 200 None run container_id:roottestchecktable-gw4-node1-1 detach:False nothrow:False cmd: ['bash', '-c', 'rm -r /var/lib/clickhouse/store/54d/54d7754d-6fdc-4e26-9c19-04b7d340750a/201902_1_1_0//checksums.txt'] Command:[docker exec --privileged roottestchecktable-gw4-node1-1 bash -c rm -r /var/lib/clickhouse/store/54d/54d7754d-6fdc-4e26-9c19-04b7d340750a/201902_1_1_0//checksums.txt] Executing query SELECT count(*) FROM numbers_mt(10000000) SETTINGS use_concurrency_control = 0 on node2 [gw2] PASSED test_concurrent_threads_soft_limit/test.py::test_use_concurrency_control_default test_concurrent_threads_soft_limit/test.py::test_use_concurrency_control_soft_limit_defined_50 Stdout:clientA3.com has address 172.16.9.2 IP: 1: lo: mtu 65536 qdisc noqueue state UNKNOWN group default qlen 1000 link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 inet 127.0.0.1/8 scope host lo valid_lft forever preferred_lft forever inet6 ::1/128 scope host valid_lft forever preferred_lft forever 48: eth0@if49: mtu 1500 qdisc noqueue state UP group default link/ether 02:42:ac:10:09:02 brd ff:ff:ff:ff:ff:ff link-netnsid 0 inet 172.16.9.2/24 brd 172.16.9.255 scope global eth0 valid_lft forever preferred_lft forever , A1 clientA1.com has address 172.16.9.4 , A2 clientA2.com has address 172.16.9.5 , A3 clientA3.com has address 172.16.9.2 run container_id:roottestallowedclienthosts-gw8-clientA3-1 detach:False nothrow:False cmd: ['bash', '-c', "/usr/bin/clickhouse client --host clientA3.com --query 'SELECT fqdn(), hostName()'"] Command:[docker exec roottestallowedclienthosts-gw8-clientA3-1 bash -c /usr/bin/clickhouse client --host clientA3.com --query 'SELECT fqdn(), hostName()'] http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None Executing query CREATE TABLE db1.table_memory (date Date, id UInt32, value Int32) ENGINE = Memory() on node1 http://localhost:None "GET /v1.46/containers/f4288bcc543c09f5b91784408f69373fa57bbff9651125c7cf2d841bebe0085a/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None Executing query SYSTEM FLUSH LOGS on node2 Stdout:clientA3.com clientA3.com s:server server , a1:clientA1.com clientA1.com , a2:clientA2.com clientA2.com , a3:clientA3.com clientA3.com run container_id:roottestallowedclienthosts-gw8-clientA1-1 detach:False nothrow:False cmd: ['bash', '-c', '/usr/bin/curl -s server:8123 '] Command:[docker exec roottestallowedclienthosts-gw8-clientA1-1 bash -c /usr/bin/curl -s server:8123 ] http://localhost:None "GET /v1.46/containers/f4288bcc543c09f5b91784408f69373fa57bbff9651125c7cf2d841bebe0085a/json HTTP/1.1" 200 None ClickHouse node1 started get_instance_ip instance_name=node2 Executing query INSERT INTO db1.table_memory VALUES (toDate('2019-02-01'), 1, 10), (toDate('2019-02-01'), 2, 12) on node1 http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw3-node2-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node2 http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw3-node2-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node2, ip: 172.16.2.8... http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw3-node2-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/79310bd7f99bac53f29b42a779f9c7a19ccc83d24e3a8dae8de5be6270acde3b/json HTTP/1.1" 200 None ClickHouse node2 started get_instance_ip instance_name=node3 http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw3-node3-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node3 http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw3-node3-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node3, ip: 172.16.2.6... http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw3-node3-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/0c68be5a1acc0d25e3a0f8b3e11b337aab4548ceb3bad9ce6fc73d93ecb72cd8/json HTTP/1.1" 200 None ClickHouse node3 started get_instance_ip instance_name=node4 http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw3-node4-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node4 http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw3-node4-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node4, ip: 172.16.2.7... http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw3-node4-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/9bd6ceee3cd6db6688a84c6f68e3fed164e8daa79c8cecbc2e33b89aabd88b62/json HTTP/1.1" 200 None ClickHouse node4 started get_instance_ip instance_name=node5 http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw3-node5-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node5 http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw3-node5-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node5, ip: 172.16.2.10... http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw3-node5-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/ee72952982b8089a80c5c7f81fd1be95d9119fa565bac23fce52778c8f50a659/json HTTP/1.1" 200 None Stdout:Ok. run container_id:roottestallowedclienthosts-gw8-clientA1-1 detach:False nothrow:False cmd: ['ip', 'address', 'show'] Command:[docker exec roottestallowedclienthosts-gw8-clientA1-1 ip address show] Stderr: Container roottestbackuprestoreoncluster-gw0-zoo1-1 Running Stderr: Container roottestbackuprestoreoncluster-gw0-zoo2-1 Running Stderr: Container roottestbackuprestoreoncluster-gw0-zoo3-1 Running Stderr: Container roottestbackuprestoreoncluster-gw0-node1-1 Creating Stderr: Container roottestbackuprestoreoncluster-gw0-node2-1 Creating Stderr: Container roottestbackuprestoreoncluster-gw0-node3-1 Creating Stderr: Container roottestbackuprestoreoncluster-gw0-node3-1 Created Stderr: Container roottestbackuprestoreoncluster-gw0-node1-1 Created Stderr: Container roottestbackuprestoreoncluster-gw0-node2-1 Created Stderr: Container roottestbackuprestoreoncluster-gw0-node3-1 Starting Stderr: Container roottestbackuprestoreoncluster-gw0-node1-1 Starting Stderr: Container roottestbackuprestoreoncluster-gw0-node2-1 Starting Stderr: Container roottestbackuprestoreoncluster-gw0-node3-1 Started Stderr: Container roottestbackuprestoreoncluster-gw0-node1-1 Started Stderr: Container roottestbackuprestoreoncluster-gw0-node2-1 Started ClickHouse instance created get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None Executing query select ProfileEvents['ConcurrencyControlSlotsGranted'] from system.query_log where current_database = currentDatabase() and query_id = 'test_use_concurrency_control_2' and type = 'QueryFinish' order by query_start_time_microseconds desc limit 1 on node2 http://localhost:None "GET /v1.46/containers/roottestbackuprestoreoncluster-gw0-node1-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottestbackuprestoreoncluster-gw0-node1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/ee72952982b8089a80c5c7f81fd1be95d9119fa565bac23fce52778c8f50a659/json HTTP/1.1" 200 None Waiting for ClickHouse start in node1, ip: 172.16.10.6... ClickHouse node5 started get_instance_ip instance_name=node6 http://localhost:None "GET /v1.46/containers/roottestbackuprestoreoncluster-gw0-node1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw3-node6-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/9667a4a5b389cc832a96cad0e61e71319228023fedeb66aa020559aaab318041/json HTTP/1.1" 200 None get_instance_ip instance_name=node6 http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw3-node6-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node6, ip: 172.16.2.9... http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw3-node6-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/bb77bcdeb206dae3948087a8335fad2b3d01e56f46dafd3666dd22a31259c749/json HTTP/1.1" 200 None ClickHouse node6 started get_instance_ip instance_name=node7 http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw3-node7-1/json HTTP/1.1" 200 None Executing query SYSTEM STOP MERGES db1.table_memory on node1 get_instance_ip instance_name=node7 http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw3-node7-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node7, ip: 172.16.2.11... http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw3-node7-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/003091bcacf6694e00f2bb1065e6688d6a27cd4492c46823d4c326c678df7c34/json HTTP/1.1" 200 None ClickHouse node7 started get_instance_ip instance_name=node8 http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw3-node8-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node8 http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw3-node8-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node8, ip: 172.16.2.5... http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw3-node8-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/3c6f23fd4a82d81438a3c36511cb5150add83da41f4c50992a71d5aa3b1fe6c8/json HTTP/1.1" 200 None ClickHouse node8 started Executing query DROP DATABASE IF EXISTS db1 SYNC on node5 Stdout:1: lo: mtu 65536 qdisc noqueue state UNKNOWN group default qlen 1000 Stdout: link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 Stdout: inet 127.0.0.1/8 scope host lo Stdout: valid_lft forever preferred_lft forever Stdout: inet6 ::1/128 scope host Stdout: valid_lft forever preferred_lft forever Stdout:53: eth0@if54: mtu 1500 qdisc noqueue state UP group default Stdout: link/ether 02:42:ac:10:09:04 brd ff:ff:ff:ff:ff:ff link-netnsid 0 Stdout: inet 172.16.9.4/24 brd 172.16.9.255 scope global eth0 Executing query CREATE TABLE db1.table_log (date Date, id UInt32, value Int32) ENGINE = TinyLog() on node1 Stdout: valid_lft forever preferred_lft forever run container_id:roottestallowedclienthosts-gw8-clientA1-1 detach:False nothrow:False cmd: ['host', 'clientA1.com'] Command:[docker exec roottestallowedclienthosts-gw8-clientA1-1 host clientA1.com] http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/9667a4a5b389cc832a96cad0e61e71319228023fedeb66aa020559aaab318041/json HTTP/1.1" 200 None Executing query CREATE DATABASE db1 on node5 Executing query select ProfileEvents['ConcurrencyControlSlotsDelayed'] from system.query_log where current_database = currentDatabase() and query_id = 'test_use_concurrency_control_2' and type = 'QueryFinish' order by query_start_time_microseconds desc limit 1 on node2 Stderr: Container roottestaltercodec-gw9-node1-1 Stopping Stderr: Container roottestaltercodec-gw9-node1-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_alter_codec/_instances-0-gw9/node1/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_alter_codec/_instances-0-gw9/node1/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Executing query INSERT INTO db1.table_log VALUES (toDate('2019-02-01'), 1, 10), (toDate('2019-02-01'), 2, 12) on node1 Command:[docker compose --env-file /ClickHouse/tests/integration/test_alter_codec/_instances-0-gw9/.env --project-name roottestaltercodec-gw9 --file /ClickHouse/tests/integration/test_alter_codec/_instances-0-gw9/node1/docker-compose.yml down --volumes] Stdout:clientA1.com has address 172.16.9.4 run container_id:roottestallowedclienthosts-gw8-clientA1-1 detach:False nothrow:False cmd: ['host', 'clientA2.com'] Command:[docker exec roottestallowedclienthosts-gw8-clientA1-1 host clientA2.com] Executing query CREATE TABLE db1.test_table(date Date, k1 String, v1 Int32) ENGINE = MergeTree(date, (k1, date), 8192) on node5 http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/9667a4a5b389cc832a96cad0e61e71319228023fedeb66aa020559aaab318041/json HTTP/1.1" 200 None Executing query select ProfileEvents['ConcurrencyControlSlotsAcquired'] from system.query_log where current_database = currentDatabase() and query_id = 'test_use_concurrency_control_2' and type = 'QueryFinish' order by query_start_time_microseconds desc limit 1 on node2 Executing query INSERT INTO db1.test_table VALUES('2000-01-01', 'test_key', 1) on node5 Executing query SYSTEM STOP MERGES db1.table_log on node1 Stdout:clientA2.com has address 172.16.9.5 run container_id:roottestallowedclienthosts-gw8-clientA1-1 detach:False nothrow:False cmd: ['host', 'clientA3.com'] Command:[docker exec roottestallowedclienthosts-gw8-clientA1-1 host clientA3.com] Executing query SYSTEM ENABLE FAILPOINT check_table_query_delay_for_part on node1 Executing query select ProfileEvents['ConcurrencyControlSlotsAcquiredNonCompeting'] from system.query_log where current_database = currentDatabase() and query_id = 'test_use_concurrency_control_2' and type = 'QueryFinish' order by query_start_time_microseconds desc limit 1 on node2 Executing query SELECT name FROM system.databases WHERE name = 'db1' on node5 http://localhost:None "GET /v1.46/containers/9667a4a5b389cc832a96cad0e61e71319228023fedeb66aa020559aaab318041/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None Executing query CHECK ALL TABLES on node1 Executing query DROP TABLE db3.table2 on node1 Executing query DROP DATABASE db2 on node1 Stdout:clientA3.com has address 172.16.9.2 IP: 1: lo: mtu 65536 qdisc noqueue state UNKNOWN group default qlen 1000 link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 inet 127.0.0.1/8 scope host lo valid_lft forever preferred_lft forever inet6 ::1/128 scope host valid_lft forever preferred_lft forever 53: eth0@if54: mtu 1500 qdisc noqueue state UP group default link/ether 02:42:ac:10:09:04 brd ff:ff:ff:ff:ff:ff link-netnsid 0 inet 172.16.9.4/24 brd 172.16.9.255 scope global eth0 valid_lft forever preferred_lft forever , A1 clientA1.com has address 172.16.9.4 , A2 clientA2.com has address 172.16.9.5 , A3 clientA3.com has address 172.16.9.2 run container_id:roottestallowedclienthosts-gw8-clientA1-1 detach:False nothrow:False cmd: ['bash', '-c', "/usr/bin/clickhouse client --host server --query 'SELECT * FROM test_allowed_client_hosts'"] Command:[docker exec roottestallowedclienthosts-gw8-clientA1-1 bash -c /usr/bin/clickhouse client --host server --query 'SELECT * FROM test_allowed_client_hosts'] Executing query SELECT name FROM system.tables WHERE database = 'db1' AND name = 'test_table' on node5 Executing query select ProfileEvents['ConcurrencyControlQueriesDelayed'] from system.query_log where current_database = currentDatabase() and query_id = 'test_use_concurrency_control_2' and type = 'QueryFinish' order by query_start_time_microseconds desc limit 1 on node2 http://localhost:None "GET /v1.46/containers/9667a4a5b389cc832a96cad0e61e71319228023fedeb66aa020559aaab318041/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None Command:[docker compose --env-file /ClickHouse/tests/integration/test_concurrent_threads_soft_limit/_instances-0-gw2/.env --project-name roottestconcurrentthreadssoftlimit-gw2 --file /ClickHouse/tests/integration/test_concurrent_threads_soft_limit/_instances-0-gw2/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_concurrent_threads_soft_limit/_instances-0-gw2/node2/docker-compose.yml --file /ClickHouse/tests/integration/test_concurrent_threads_soft_limit/_instances-0-gw2/node3/docker-compose.yml --file /ClickHouse/tests/integration/test_concurrent_threads_soft_limit/_instances-0-gw2/node4/docker-compose.yml stop --timeout 20] Executing query SELECT name FROM system.columns WHERE database = 'db1' AND table = 'test_table' on node5 [gw2] PASSED test_concurrent_threads_soft_limit/test.py::test_use_concurrency_control_soft_limit_defined_50 Stdout:5 run container_id:roottestallowedclienthosts-gw8-clientA3-1 detach:False nothrow:False cmd: ['bash', '-c', '/usr/bin/curl -s server:8123 '] Command:[docker exec roottestallowedclienthosts-gw8-clientA3-1 bash -c /usr/bin/curl -s server:8123 ] http://localhost:None "GET /v1.46/containers/9667a4a5b389cc832a96cad0e61e71319228023fedeb66aa020559aaab318041/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None Executing query SELECT name FROM system.parts WHERE database = 'db1' AND table = 'test_table' on node5 Executing query SELECT name FROM system.parts_columns WHERE database = 'db1' AND table = 'test_table' on node5 Stdout:Ok. run container_id:roottestallowedclienthosts-gw8-clientA3-1 detach:False nothrow:False cmd: ['ip', 'address', 'show'] Command:[docker exec roottestallowedclienthosts-gw8-clientA3-1 ip address show] http://localhost:None "GET /v1.46/containers/9667a4a5b389cc832a96cad0e61e71319228023fedeb66aa020559aaab318041/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None run container_id:roottestaccessforfunctions-gw7-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestaccessforfunctions-gw7-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:1: lo: mtu 65536 qdisc noqueue state UNKNOWN group default qlen 1000 Stdout: link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 Stdout: inet 127.0.0.1/8 scope host lo Stdout: valid_lft forever preferred_lft forever Stdout: inet6 ::1/128 scope host Stdout: valid_lft forever preferred_lft forever Stdout:48: eth0@if49: mtu 1500 qdisc noqueue state UP group default Stdout: link/ether 02:42:ac:10:09:02 brd ff:ff:ff:ff:ff:ff link-netnsid 0 Stdout: inet 172.16.9.2/24 brd 172.16.9.255 scope global eth0 Stdout: valid_lft forever preferred_lft forever run container_id:roottestallowedclienthosts-gw8-clientA3-1 detach:False nothrow:False cmd: ['host', 'clientA1.com'] Command:[docker exec roottestallowedclienthosts-gw8-clientA3-1 host clientA1.com] Executing query SELECT name FROM system.databases WHERE name = 'db1' on node5 Stdout:768 Clickhouse process running. run container_id:roottestaccessforfunctions-gw7-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestaccessforfunctions-gw7-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] http://localhost:None "GET /v1.46/containers/9667a4a5b389cc832a96cad0e61e71319228023fedeb66aa020559aaab318041/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None Stdout:clientA1.com has address 172.16.9.4 run container_id:roottestallowedclienthosts-gw8-clientA3-1 detach:False nothrow:False cmd: ['host', 'clientA2.com'] Command:[docker exec roottestallowedclienthosts-gw8-clientA3-1 host clientA2.com] Stdout:768 Executing query select 20 on instance Executing query SELECT name FROM system.tables WHERE database = 'db1' AND name = 'test_table' on node5 Executing query SHOW GRANTS FOR `d481e934-6745-42eb-86e7-1d8aec307361` on instance Stderr: Container roottestaltercodec-gw9-node1-1 Stopping Stderr: Container roottestaltercodec-gw9-node1-1 Stopped Stderr: Container roottestaltercodec-gw9-node1-1 Removing Stderr: Container roottestaltercodec-gw9-node1-1 Removed Stderr: Network roottestaltercodec-gw9_default Removing Stderr: Network roottestaltercodec-gw9_default Removed Cleanup called http://localhost:None "GET /v1.46/containers/9667a4a5b389cc832a96cad0e61e71319228023fedeb66aa020559aaab318041/json HTTP/1.1" 200 None ClickHouse node1 started Docker networks for project roottestaltercodec-gw9 are NETWORK ID NAME DRIVER SCOPE get_instance_ip instance_name=node2 Stdout:clientA2.com has address 172.16.9.5 run container_id:roottestallowedclienthosts-gw8-clientA3-1 detach:False nothrow:False cmd: ['host', 'clientA3.com'] Command:[docker exec roottestallowedclienthosts-gw8-clientA3-1 host clientA3.com] http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestbackuprestoreoncluster-gw0-node2-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node2 http://localhost:None "GET /v1.46/containers/roottestbackuprestoreoncluster-gw0-node2-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node2, ip: 172.16.10.5... http://localhost:None "GET /v1.46/containers/roottestbackuprestoreoncluster-gw0-node2-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/62b40abd527625176ef363ace367e3ed09c301016410775e58d0ec2a1cc62785/json HTTP/1.1" 200 None ClickHouse node2 started get_instance_ip instance_name=node3 http://localhost:None "GET /v1.46/containers/roottestbackuprestoreoncluster-gw0-node3-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node3 http://localhost:None "GET /v1.46/containers/roottestbackuprestoreoncluster-gw0-node3-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node3, ip: 172.16.10.7... http://localhost:None "GET /v1.46/containers/roottestbackuprestoreoncluster-gw0-node3-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/0f4702ec03365f9c6180c49cad250532b772a2c52121c5c434267dc2bc7419f9/json HTTP/1.1" 200 None ClickHouse node3 started Docker containers for project roottestaltercodec-gw9 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Executing query CREATE TABLE tbl ON CLUSTER 'cluster' (x UInt8) ENGINE=ReplicatedMergeTree('/clickhouse/tables/tbl/', '{replica}')ORDER BY x on node1 Command:[docker compose --env-file /ClickHouse/tests/integration/test_access_for_functions/_instances-0-gw7/.env --project-name roottestaccessforfunctions-gw7 --file /ClickHouse/tests/integration/test_access_for_functions/_instances-0-gw7/instance/docker-compose.yml stop --timeout 20] [gw7] PASSED test_access_for_functions/test.py::test_ignore_obsolete_grant_on_database Executing query SELECT name FROM system.columns WHERE database = 'db1' AND table = 'test_table' on node5 Docker volumes for project roottestaltercodec-gw9 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestaltercodec-gw9-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestaltercodec-gw9 Trying to prune unused networks... Stdout:clientA3.com has address 172.16.9.2 IP: 1: lo: mtu 65536 qdisc noqueue state UNKNOWN group default qlen 1000 link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 inet 127.0.0.1/8 scope host lo valid_lft forever preferred_lft forever inet6 ::1/128 scope host valid_lft forever preferred_lft forever 48: eth0@if49: mtu 1500 qdisc noqueue state UP group default link/ether 02:42:ac:10:09:02 brd ff:ff:ff:ff:ff:ff link-netnsid 0 inet 172.16.9.2/24 brd 172.16.9.255 scope global eth0 valid_lft forever preferred_lft forever , A1 clientA1.com has address 172.16.9.4 , A2 clientA2.com has address 172.16.9.5 , A3 clientA3.com has address 172.16.9.2 run container_id:roottestallowedclienthosts-gw8-clientA3-1 detach:False nothrow:False cmd: ['bash', '-c', "/usr/bin/clickhouse client --host server --query 'SELECT * FROM test_allowed_client_hosts'"] Command:[docker exec roottestallowedclienthosts-gw8-clientA3-1 bash -c /usr/bin/clickhouse client --host server --query 'SELECT * FROM test_allowed_client_hosts'] Trying to prune unused images... Command:[docker image prune -f] http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None Executing query SELECT name FROM system.parts WHERE database = 'db1' AND table = 'test_table' on node5 Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:1 Volumes pruned: 1 Running tests in /ClickHouse/tests/integration/test_attach_table_normalizer/test.py test_attach_table_normalizer/test.py::test_attach_substr Cluster start called. is_up=False Docker networks for project roottestattachtablenormalizer-gw9 are NETWORK ID NAME DRIVER SCOPE Executing query SELECT name FROM system.parts_columns WHERE database = 'db1' AND table = 'test_table' on node5 Stdout:5 [gw8] PASSED test_allowed_client_hosts/test.py::test_allowed_host run container_id:roottestallowedclienthosts-gw8-clientA2-1 detach:False nothrow:False cmd: ['bash', '-c', '/usr/bin/curl -s server:8123 '] Command:[docker exec roottestallowedclienthosts-gw8-clientA2-1 bash -c /usr/bin/curl -s server:8123 ] test_allowed_client_hosts/test.py::test_denied_host Docker containers for project roottestattachtablenormalizer-gw9 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None Docker volumes for project roottestattachtablenormalizer-gw9 are DRIVER VOLUME NAME Cleanup called Stdout:Ok. run container_id:roottestallowedclienthosts-gw8-clientA2-1 detach:False nothrow:False cmd: ['ip', 'address', 'show'] Command:[docker exec roottestallowedclienthosts-gw8-clientA2-1 ip address show] Executing query INSERT INTO tbl VALUES (1) on node1 [gw3] PASSED test_config_substitutions/test.py::test_allow_databases Executing query select value from system.settings where name = 'max_query_size' on node1 test_config_substitutions/test.py::test_config Docker networks for project roottestattachtablenormalizer-gw9 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestattachtablenormalizer-gw9 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestattachtablenormalizer-gw9 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestattachtablenormalizer-gw9-.*-1$' --format '{{.ID}}:{{.Names}}'] Executing query BACKUP TABLE tbl TO Disk('backups', '1') ASYNC on node1 via HTTP interface Starting new HTTP connection (1): 172.16.10.6:8123 http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None Stdout:1: lo: mtu 65536 qdisc noqueue state UNKNOWN group default qlen 1000 Stdout: link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 Stdout: inet 127.0.0.1/8 scope host lo Stdout: valid_lft forever preferred_lft forever Stdout: inet6 ::1/128 scope host Stdout: valid_lft forever preferred_lft forever Stdout:55: eth0@if56: mtu 1500 qdisc noqueue state UP group default Stdout: link/ether 02:42:ac:10:09:05 brd ff:ff:ff:ff:ff:ff link-netnsid 0 Stdout: inet 172.16.9.5/24 brd 172.16.9.255 scope global eth0 Stdout: valid_lft forever preferred_lft forever run container_id:roottestallowedclienthosts-gw8-clientA2-1 detach:False nothrow:False cmd: ['host', 'clientA1.com'] Command:[docker exec roottestallowedclienthosts-gw8-clientA2-1 host clientA1.com] http://172.16.10.6:8123 "GET /?query=BACKUP+TABLE+tbl++TO+Disk%28%27backups%27%2C+%271%27%29+ASYNC HTTP/1.1" 200 None Unstopped containers: {} No running containers for project: roottestattachtablenormalizer-gw9 Trying to prune unused networks... Executing query BACKUP TABLE tbl TO Disk('backups', '1') ASYNC on node2 via HTTP interface Starting new HTTP connection (1): 172.16.10.5:8123 http://172.16.10.5:8123 "GET /?query=BACKUP+TABLE+tbl++TO+Disk%28%27backups%27%2C+%271%27%29+ASYNC HTTP/1.1" 200 None Executing query SELECT status FROM system.backups WHERE id IN ['1949b493-f79e-47df-a3ae-ec8c510bc3fe', '49d6d210-3941-463a-a409-84c63ab2973b'] AND status == 'CREATING_BACKUP' on node1 Trying to prune unused images... Command:[docker image prune -f] Executing query select value from system.settings where name = 'max_query_size' on node2 Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:clientA1.com has address 172.16.9.4 run container_id:roottestallowedclienthosts-gw8-clientA2-1 detach:False nothrow:False cmd: ['host', 'clientA2.com'] Command:[docker exec roottestallowedclienthosts-gw8-clientA2-1 host clientA2.com] Stdout:1 Volumes pruned: 1 Setup directory for instance: node Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files [] to /ClickHouse/tests/integration/test_attach_table_normalizer/_instances-0-gw9/node/configs/config.d Setup database dir /ClickHouse/tests/integration/test_attach_table_normalizer/_instances-0-gw9/node/database Setup logs dir /ClickHouse/tests/integration/test_attach_table_normalizer/_instances-0-gw9/node/logs Entrypoint cmd: bash -c "trap 'pkill tail' INT TERM; clickhouse server --config-file=/etc/clickhouse-server/config.xml --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log --daemon -- ; coproc tail -f /dev/null; wait $$!" Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw'} stored in /ClickHouse/tests/integration/test_attach_table_normalizer/_instances-0-gw9/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_attach_table_normalizer/_instances-0-gw9/.env --project-name roottestattachtablenormalizer-gw9 --file /ClickHouse/tests/integration/test_attach_table_normalizer/_instances-0-gw9/node/docker-compose.yml pull] http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None Executing query SELECT * FROM test.tbl2 ORDER BY p on node Executing query select value from system.settings where name = 'max_query_size' on node3 Stdout:clientA2.com has address 172.16.9.5 run container_id:roottestallowedclienthosts-gw8-clientA2-1 detach:False nothrow:False cmd: ['host', 'clientA3.com'] Command:[docker exec roottestallowedclienthosts-gw8-clientA2-1 host clientA3.com] Executing query SELECT status FROM system.backups WHERE id IN ['1949b493-f79e-47df-a3ae-ec8c510bc3fe', '49d6d210-3941-463a-a409-84c63ab2973b'] AND status == 'CREATING_BACKUP' on node2 http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None Stderr: Container roottestaccessforfunctions-gw7-instance-1 Stopping Stderr: Container roottestaccessforfunctions-gw7-instance-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_access_for_functions/_instances-0-gw7/instance/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_access_for_functions/_instances-0-gw7/instance/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Executing query select value from system.settings where name = 'max_query_size' on node4 Command:[docker compose --env-file /ClickHouse/tests/integration/test_access_for_functions/_instances-0-gw7/.env --project-name roottestaccessforfunctions-gw7 --file /ClickHouse/tests/integration/test_access_for_functions/_instances-0-gw7/instance/docker-compose.yml down --volumes] Executing query DROP TABLE IF EXISTS test.tbl2 on node Stdout:clientA3.com has address 172.16.9.2 IP: 1: lo: mtu 65536 qdisc noqueue state UNKNOWN group default qlen 1000 link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 inet 127.0.0.1/8 scope host lo valid_lft forever preferred_lft forever inet6 ::1/128 scope host valid_lft forever preferred_lft forever 55: eth0@if56: mtu 1500 qdisc noqueue state UP group default link/ether 02:42:ac:10:09:05 brd ff:ff:ff:ff:ff:ff link-netnsid 0 inet 172.16.9.5/24 brd 172.16.9.255 scope global eth0 valid_lft forever preferred_lft forever , A1 clientA1.com has address 172.16.9.4 , A2 clientA2.com has address 172.16.9.5 , A3 clientA3.com has address 172.16.9.2 run container_id:roottestallowedclienthosts-gw8-clientA2-1 detach:False nothrow:False cmd: ['bash', '-c', "/usr/bin/clickhouse client --host server --query 'SELECT * FROM test_allowed_client_hosts'"] Command:[docker exec roottestallowedclienthosts-gw8-clientA2-1 bash -c /usr/bin/clickhouse client --host server --query 'SELECT * FROM test_allowed_client_hosts'] http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None Executing query select value from system.settings where name = 'max_query_size' on node6 Executing query SELECT count() FROM system.backups WHERE id IN ['1949b493-f79e-47df-a3ae-ec8c510bc3fe', '49d6d210-3941-463a-a409-84c63ab2973b'] AND status == 'BACKUP_CREATED' on node1 [gw5] PASSED test_backup_restore/test.py::test_attach_partition Executing query CREATE TABLE test.tbl3 AS test.tbl on node test_backup_restore/test.py::test_replace_partition Executing query select value from system.settings where name = 'max_query_size' on node7 http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None Stderr:Code: 194. DB::Exception: Received from server:9000. DB::Exception: default: Authentication failed: password is incorrect, or there is no user with such name. Stderr: Stderr:If you use ClickHouse Cloud, the password can be reset at https://clickhouse.cloud/ Stderr:on the settings page for the corresponding service. Stderr: Stderr:If you have installed ClickHouse and forgot password you can reset it in the configuration file. Stderr:The password for default user is typically located at /etc/clickhouse-server/users.d/default-password.xml Stderr:and deleting this file will reset the password. Stderr:See also /etc/clickhouse-server/users.xml on the server where ClickHouse is installed. Stderr: Stderr:. (REQUIRED_PASSWORD) Stderr: Exitcode:194 Command:[docker compose --env-file /ClickHouse/tests/integration/test_allowed_client_hosts/_instances-0-gw8/.env --project-name roottestallowedclienthosts-gw8 --file /ClickHouse/tests/integration/test_allowed_client_hosts/_instances-0-gw8/server/docker-compose.yml --file /ClickHouse/tests/integration/test_allowed_client_hosts/_instances-0-gw8/clientA1/docker-compose.yml --file /ClickHouse/tests/integration/test_allowed_client_hosts/_instances-0-gw8/clientA2/docker-compose.yml --file /ClickHouse/tests/integration/test_allowed_client_hosts/_instances-0-gw8/clientA3/docker-compose.yml stop --timeout 20] [gw8] PASSED test_allowed_client_hosts/test.py::test_denied_host Executing query SELECT count() FROM system.backups WHERE id IN ['1949b493-f79e-47df-a3ae-ec8c510bc3fe', '49d6d210-3941-463a-a409-84c63ab2973b'] AND status == 'BACKUP_CREATED' on node2 Executing query select value from system.settings where name = 'max_threads' on node7 Executing query INSERT INTO test.tbl3(p, k) VALUES(toDate(3), 3) on node http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None Executing query select value from system.settings where name = 'max_query_size' on node8 Executing query INSERT INTO test.tbl3(p, k) VALUES(toDate(4), 4) on node Executing query SELECT count() FROM system.backups WHERE id IN ['1949b493-f79e-47df-a3ae-ec8c510bc3fe', '49d6d210-3941-463a-a409-84c63ab2973b'] AND status == 'BACKUP_FAILED' on node1 http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None [gw3] PASSED test_config_substitutions/test.py::test_config run container_id:roottestconfigsubstitutions-gw3-node7-1 detach:False nothrow:False cmd: ['bash', '-c', 'test ! -f /etc/clickhouse-server/users.d/000-users_with_env_subst.xml || mv --no-clobber /etc/clickhouse-server/users.d/000-users_with_env_subst.xml /tmp/000-users_with_env_subst_2b14efe8-341d-4703-9033-3b137deb666c.xml'] test_config_substitutions/test.py::test_config_from_env_overrides Command:[docker exec roottestconfigsubstitutions-gw3-node7-1 bash -c test ! -f /etc/clickhouse-server/users.d/000-users_with_env_subst.xml || mv --no-clobber /etc/clickhouse-server/users.d/000-users_with_env_subst.xml /tmp/000-users_with_env_subst_2b14efe8-341d-4703-9033-3b137deb666c.xml] Executing query INSERT INTO test.tbl3(p, k) VALUES(toDate(33), 33) on node Executing query SELECT count() FROM system.backups WHERE id IN ['1949b493-f79e-47df-a3ae-ec8c510bc3fe', '49d6d210-3941-463a-a409-84c63ab2973b'] AND status == 'BACKUP_FAILED' on node2 run container_id:roottestconfigsubstitutions-gw3-node7-1 detach:False nothrow:False cmd: ['bash', '-c', 'echo \'\n\n \n \n \n 100\n \n \n \n \n \n default\n default\n \n\n \n \n \n\n\' > /etc/clickhouse-server/users.d/000-users_with_env_subst.xml'] Command:[docker exec roottestconfigsubstitutions-gw3-node7-1 bash -c echo ' 100 default default ' > /etc/clickhouse-server/users.d/000-users_with_env_subst.xml] Executing query INSERT INTO test.tbl3(p, k) VALUES(toDate(34), 34) on node http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None Executing query DROP TABLE tbl ON CLUSTER 'cluster' SYNC on node1 Executing query SYSTEM RELOAD CONFIG on node7 Executing query SELECT * FROM test.tbl3 ORDER BY p on node http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None copy from /var/lib/clickhouse/shadow/*/data/test/tbl to /var/lib/clickhouse/data/test/tbl3/detached run container_id:roottestbackuprestore-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', 'cp -r /var/lib/clickhouse/shadow/*/data/test/tbl -T /var/lib/clickhouse/data/test/tbl3/detached'] Command:[docker exec roottestbackuprestore-gw5-node-1 bash -c cp -r /var/lib/clickhouse/shadow/*/data/test/tbl -T /var/lib/clickhouse/data/test/tbl3/detached] run container_id:roottestconfigsubstitutions-gw3-node7-1 detach:False nothrow:False cmd: ['bash', '-c', 'test ! -f /tmp/000-users_with_env_subst_2b14efe8-341d-4703-9033-3b137deb666c.xml || mv /tmp/000-users_with_env_subst_2b14efe8-341d-4703-9033-3b137deb666c.xml /etc/clickhouse-server/users.d/000-users_with_env_subst.xml'] Command:[docker exec roottestconfigsubstitutions-gw3-node7-1 bash -c test ! -f /tmp/000-users_with_env_subst_2b14efe8-341d-4703-9033-3b137deb666c.xml || mv /tmp/000-users_with_env_subst_2b14efe8-341d-4703-9033-3b137deb666c.xml /etc/clickhouse-server/users.d/000-users_with_env_subst.xml] Executing query RESTORE TABLE tbl FROM Disk('backups', '1') on node1 Stderr: Container roottestaccessforfunctions-gw7-instance-1 Stopping Stderr: Container roottestaccessforfunctions-gw7-instance-1 Stopped Stderr: Container roottestaccessforfunctions-gw7-instance-1 Removing Stderr: Container roottestaccessforfunctions-gw7-instance-1 Removed Stderr: Network roottestaccessforfunctions-gw7_default Removing Stderr: Network roottestaccessforfunctions-gw7_default Removed Cleanup called Executing query ALTER TABLE test.tbl3 REPLACE PARTITION 197002 FROM test.tbl on node run container_id:roottestconfigsubstitutions-gw3-node7-1 detach:False nothrow:False cmd: ['bash', '-c', 'test ! -f /etc/clickhouse-server/users.d/000-users_with_env_subst.xml || mv --no-clobber /etc/clickhouse-server/users.d/000-users_with_env_subst.xml /tmp/000-users_with_env_subst_77d71341-539a-4ba1-aa40-6d0128ccad10.xml'] Command:[docker exec roottestconfigsubstitutions-gw3-node7-1 bash -c test ! -f /etc/clickhouse-server/users.d/000-users_with_env_subst.xml || mv --no-clobber /etc/clickhouse-server/users.d/000-users_with_env_subst.xml /tmp/000-users_with_env_subst_77d71341-539a-4ba1-aa40-6d0128ccad10.xml] http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None Docker networks for project roottestaccessforfunctions-gw7 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestaccessforfunctions-gw7 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Executing query SELECT sleep(2) on node Docker volumes for project roottestaccessforfunctions-gw7 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestaccessforfunctions-gw7-.*-1$' --format '{{.ID}}:{{.Names}}'] run container_id:roottestconfigsubstitutions-gw3-node7-1 detach:False nothrow:False cmd: ['bash', '-c', 'echo \'\n\n \n \n \n 1\n \n \n \n \n \n default\n default\n \n\n \n \n \n\n\' > /etc/clickhouse-server/users.d/000-users_with_env_subst.xml'] Command:[docker exec roottestconfigsubstitutions-gw3-node7-1 bash -c echo ' 1 default default ' > /etc/clickhouse-server/users.d/000-users_with_env_subst.xml] Executing query SELECT * FROM tbl on node1 Unstopped containers: {} No running containers for project: roottestaccessforfunctions-gw7 Trying to prune unused networks... http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Executing query SYSTEM RELOAD CONFIG on node7 Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw0] PASSED test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-False] Stdout:1 Volumes pruned: 1 test_alter_on_mixed_type_cluster/test.py::test_alter_on_cluter_non_replicated Running tests in /ClickHouse/tests/integration/test_alter_on_mixed_type_cluster/test.py Cluster start called. is_up=False Docker networks for project roottestalteronmixedtypecluster-gw7 are NETWORK ID NAME DRIVER SCOPE http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None run container_id:roottestconfigsubstitutions-gw3-node7-1 detach:False nothrow:False cmd: ['bash', '-c', 'test ! -f /tmp/000-users_with_env_subst_77d71341-539a-4ba1-aa40-6d0128ccad10.xml || mv /tmp/000-users_with_env_subst_77d71341-539a-4ba1-aa40-6d0128ccad10.xml /etc/clickhouse-server/users.d/000-users_with_env_subst.xml'] Command:[docker exec roottestconfigsubstitutions-gw3-node7-1 bash -c test ! -f /tmp/000-users_with_env_subst_77d71341-539a-4ba1-aa40-6d0128ccad10.xml || mv /tmp/000-users_with_env_subst_77d71341-539a-4ba1-aa40-6d0128ccad10.xml /etc/clickhouse-server/users.d/000-users_with_env_subst.xml] Docker containers for project roottestalteronmixedtypecluster-gw7 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestalteronmixedtypecluster-gw7 are DRIVER VOLUME NAME Cleanup called Docker networks for project roottestalteronmixedtypecluster-gw7 are NETWORK ID NAME DRIVER SCOPE Executing query SYSTEM RELOAD CONFIG on node7 [gw3] PASSED test_config_substitutions/test.py::test_config_from_env_overrides test_config_substitutions/test.py::test_config_merge_from_env_overrides Docker containers for project roottestalteronmixedtypecluster-gw7 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestalteronmixedtypecluster-gw7 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestalteronmixedtypecluster-gw7-.*-1$' --format '{{.ID}}:{{.Names}}'] http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Unstopped containers: {} No running containers for project: roottestalteronmixedtypecluster-gw7 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Executing query SELECT value FROM system.server_settings WHERE name='max_thread_pool_size' on node7 Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None Stdout:1 Volumes pruned: 1 Setup directory for instance: node1 run container_id:roottestconfigsubstitutions-gw3-node7-1 detach:False nothrow:False cmd: ['bash', '-c', 'test ! -f /etc/clickhouse-server/config.d/010-server_with_env_subst.xml || mv --no-clobber /etc/clickhouse-server/config.d/010-server_with_env_subst.xml /tmp/010-server_with_env_subst_a989c0b8-d2a4-48b3-a77c-fd10714c04d4.xml'] Command:[docker exec roottestconfigsubstitutions-gw3-node7-1 bash -c test ! -f /etc/clickhouse-server/config.d/010-server_with_env_subst.xml || mv --no-clobber /etc/clickhouse-server/config.d/010-server_with_env_subst.xml /tmp/010-server_with_env_subst_a989c0b8-d2a4-48b3-a77c-fd10714c04d4.xml] Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_alter_on_mixed_type_cluster/configs/remote_servers.xml'] to /ClickHouse/tests/integration/test_alter_on_mixed_type_cluster/_instances-0-gw7/node1/configs/config.d Setup database dir /ClickHouse/tests/integration/test_alter_on_mixed_type_cluster/_instances-0-gw7/node1/database Setup logs dir /ClickHouse/tests/integration/test_alter_on_mixed_type_cluster/_instances-0-gw7/node1/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Setup directory for instance: node2 Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_alter_on_mixed_type_cluster/configs/remote_servers.xml'] to /ClickHouse/tests/integration/test_alter_on_mixed_type_cluster/_instances-0-gw7/node2/configs/config.d Setup database dir /ClickHouse/tests/integration/test_alter_on_mixed_type_cluster/_instances-0-gw7/node2/database Setup logs dir /ClickHouse/tests/integration/test_alter_on_mixed_type_cluster/_instances-0-gw7/node2/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Setup directory for instance: node3 Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_alter_on_mixed_type_cluster/configs/remote_servers.xml'] to /ClickHouse/tests/integration/test_alter_on_mixed_type_cluster/_instances-0-gw7/node3/configs/config.d Setup database dir /ClickHouse/tests/integration/test_alter_on_mixed_type_cluster/_instances-0-gw7/node3/database Setup logs dir /ClickHouse/tests/integration/test_alter_on_mixed_type_cluster/_instances-0-gw7/node3/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Setup directory for instance: node4 Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_alter_on_mixed_type_cluster/configs/remote_servers.xml'] to /ClickHouse/tests/integration/test_alter_on_mixed_type_cluster/_instances-0-gw7/node4/configs/config.d Setup database dir /ClickHouse/tests/integration/test_alter_on_mixed_type_cluster/_instances-0-gw7/node4/database Setup logs dir /ClickHouse/tests/integration/test_alter_on_mixed_type_cluster/_instances-0-gw7/node4/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'keeper_binary': '/clickhouse', 'keeper_cmd_prefix': 'clickhouse keeper', 'image': 'altinityinfra/integration-test:8b2301119731', 'user': '0', 'keeper_fs': 'bind', 'keeper_logs_dir1': '/ClickHouse/tests/integration/test_alter_on_mixed_type_cluster/_instances-0-gw7/keeper1/log', 'keeper_config_dir1': '/ClickHouse/tests/integration/test_alter_on_mixed_type_cluster/_instances-0-gw7/keeper1/config', 'keeper_db_dir1': '/ClickHouse/tests/integration/test_alter_on_mixed_type_cluster/_instances-0-gw7/keeper1/coordination', 'keeper_logs_dir2': '/ClickHouse/tests/integration/test_alter_on_mixed_type_cluster/_instances-0-gw7/keeper2/log', 'keeper_config_dir2': '/ClickHouse/tests/integration/test_alter_on_mixed_type_cluster/_instances-0-gw7/keeper2/config', 'keeper_db_dir2': '/ClickHouse/tests/integration/test_alter_on_mixed_type_cluster/_instances-0-gw7/keeper2/coordination', 'keeper_logs_dir3': '/ClickHouse/tests/integration/test_alter_on_mixed_type_cluster/_instances-0-gw7/keeper3/log', 'keeper_config_dir3': '/ClickHouse/tests/integration/test_alter_on_mixed_type_cluster/_instances-0-gw7/keeper3/config', 'keeper_db_dir3': '/ClickHouse/tests/integration/test_alter_on_mixed_type_cluster/_instances-0-gw7/keeper3/coordination'} stored in /ClickHouse/tests/integration/test_alter_on_mixed_type_cluster/_instances-0-gw7/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_alter_on_mixed_type_cluster/_instances-0-gw7/.env --project-name roottestalteronmixedtypecluster-gw7 --file /ClickHouse/tests/integration/test_alter_on_mixed_type_cluster/_instances-0-gw7/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_alter_on_mixed_type_cluster/_instances-0-gw7/node2/docker-compose.yml --file /ClickHouse/tests/integration/test_alter_on_mixed_type_cluster/_instances-0-gw7/node3/docker-compose.yml --file /ClickHouse/tests/integration/test_alter_on_mixed_type_cluster/_instances-0-gw7/node4/docker-compose.yml pull] Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 run container_id:roottestconfigsubstitutions-gw3-node7-1 detach:False nothrow:False cmd: ['bash', '-c', 'echo \'\n\n 9000\n\n\' > /etc/clickhouse-server/config.d/010-server_with_env_subst.xml'] Command:[docker exec roottestconfigsubstitutions-gw3-node7-1 bash -c echo ' 9000 ' > /etc/clickhouse-server/config.d/010-server_with_env_subst.xml] http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None Executing query SYSTEM RELOAD CONFIG on node7 Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None run container_id:roottestconfigsubstitutions-gw3-node7-1 detach:False nothrow:False cmd: ['bash', '-c', 'test ! -f /tmp/010-server_with_env_subst_a989c0b8-d2a4-48b3-a77c-fd10714c04d4.xml || mv /tmp/010-server_with_env_subst_a989c0b8-d2a4-48b3-a77c-fd10714c04d4.xml /etc/clickhouse-server/config.d/010-server_with_env_subst.xml'] Command:[docker exec roottestconfigsubstitutions-gw3-node7-1 bash -c test ! -f /tmp/010-server_with_env_subst_a989c0b8-d2a4-48b3-a77c-fd10714c04d4.xml || mv /tmp/010-server_with_env_subst_a989c0b8-d2a4-48b3-a77c-fd10714c04d4.xml /etc/clickhouse-server/config.d/010-server_with_env_subst.xml] Executing query DROP TABLE IF EXISTS non_replicated_mt on node1 [gw4] PASSED test_check_table/test.py::test_check_all_tables test_check_table/test.py::test_check_normal_table_corruption[] [gw3] PASSED test_config_substitutions/test.py::test_config_merge_from_env_overrides test_config_substitutions/test.py::test_config_multiple_zk_substitutions run container_id:roottestconfigsubstitutions-gw3-node3-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] Command:[docker exec -u root roottestconfigsubstitutions-gw3-node3-1 bash -c ps -C clickhouse] Executing query CREATE TABLE non_replicated_mt(date Date, id UInt32, value Int32) ENGINE = MergeTree() PARTITION BY toYYYYMM(date) ORDER BY id ; on node1 http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Executing query INSERT INTO non_replicated_mt VALUES (toDate('2019-02-01'), 1, 10), (toDate('2019-02-01'), 2, 12) on node1 Stdout: PID TTY TIME CMD Stdout: 10 ? 00:00:00 clickhouse run container_id:roottestconfigsubstitutions-gw3-node3-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] Command:[docker exec -u root roottestconfigsubstitutions-gw3-node3-1 bash -c pkill clickhouse] http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None Executing query CHECK TABLE non_replicated_mt PARTITION 201902 on node1 run container_id:roottestconfigsubstitutions-gw3-node3-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestconfigsubstitutions-gw3-node3-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query CREATE TABLE tbl ON CLUSTER 'cluster' (x UInt8) ENGINE=ReplicatedMergeTree('/clickhouse/tables/tbl/', '{replica}')ORDER BY x on node1 test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-True] Stdout:10 http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None Executing query INSERT INTO tbl VALUES (1) on node1 http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None Executing query BACKUP TABLE tbl ON CLUSTER 'cluster' TO Disk('backups', '2') ASYNC on node1 via HTTP interface Starting new HTTP connection (1): 172.16.10.6:8123 http://172.16.10.6:8123 "GET /?query=BACKUP+TABLE+tbl+ON+CLUSTER+%27cluster%27+TO+Disk%28%27backups%27%2C+%272%27%29+ASYNC HTTP/1.1" 200 None Executing query BACKUP TABLE tbl ON CLUSTER 'cluster' TO Disk('backups', '2') ASYNC on node2 via HTTP interface Starting new HTTP connection (1): 172.16.10.5:8123 http://172.16.10.5:8123 "GET /?query=BACKUP+TABLE+tbl+ON+CLUSTER+%27cluster%27+TO+Disk%28%27backups%27%2C+%272%27%29+ASYNC HTTP/1.1" 200 None Executing query SELECT status FROM system.backups WHERE id IN ['4016971f-8a07-4fcd-ac25-177b0ad6a70e', 'ea8e213d-6bbe-498a-a117-9a151fa3b3c7'] AND status == 'CREATING_BACKUP' on node1 http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None Executing query SELECT path FROM system.parts WHERE database = 'default' AND table = 'non_replicated_mt' AND name = '201902_1_1_0' on node1 run container_id:roottestchecktable-gw4-node1-1 detach:False nothrow:False cmd: ['bash', '-c', 'rm -r /var/lib/clickhouse/store/2dd/2dd39276-fe43-4b78-b3c8-f0e157db2102/201902_1_1_0//checksums.txt'] Command:[docker exec --privileged roottestchecktable-gw4-node1-1 bash -c rm -r /var/lib/clickhouse/store/2dd/2dd39276-fe43-4b78-b3c8-f0e157db2102/201902_1_1_0//checksums.txt] http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None Executing query CHECK TABLE non_replicated_mt on node1 http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None Executing query SELECT status FROM system.backups WHERE id IN ['4016971f-8a07-4fcd-ac25-177b0ad6a70e', 'ea8e213d-6bbe-498a-a117-9a151fa3b3c7'] AND status == 'CREATING_BACKUP' on node1 http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None Executing query SELECT status FROM system.backups WHERE id IN ['4016971f-8a07-4fcd-ac25-177b0ad6a70e', 'ea8e213d-6bbe-498a-a117-9a151fa3b3c7'] AND status == 'CREATING_BACKUP' on node2 http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None run container_id:roottestconfigsubstitutions-gw3-node3-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestconfigsubstitutions-gw3-node3-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None Stdout:10 Executing query SELECT count() FROM system.backups WHERE id IN ['4016971f-8a07-4fcd-ac25-177b0ad6a70e', 'ea8e213d-6bbe-498a-a117-9a151fa3b3c7'] AND status == 'BACKUP_CREATED' on node1 http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None Executing query SELECT * FROM test.tbl3 ORDER BY p on node Executing query SELECT count() FROM system.backups WHERE id IN ['4016971f-8a07-4fcd-ac25-177b0ad6a70e', 'ea8e213d-6bbe-498a-a117-9a151fa3b3c7'] AND status == 'BACKUP_CREATED' on node2 http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None Executing query ALTER TABLE test.tbl3 UPDATE k=10 WHERE 1 on node Executing query SELECT count() FROM system.backups WHERE id IN ['4016971f-8a07-4fcd-ac25-177b0ad6a70e', 'ea8e213d-6bbe-498a-a117-9a151fa3b3c7'] AND status == 'BACKUP_FAILED' on node1 Executing query SELECT sleep(2) on node http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None Executing query SELECT count() FROM system.backups WHERE id IN ['4016971f-8a07-4fcd-ac25-177b0ad6a70e', 'ea8e213d-6bbe-498a-a117-9a151fa3b3c7'] AND status == 'BACKUP_FAILED' on node2 http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None Executing query DROP TABLE tbl ON CLUSTER 'cluster' SYNC on node1 http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None Executing query RESTORE TABLE tbl FROM Disk('backups', '2') on node1 http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None Executing query SELECT * FROM tbl on node1 http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw0] PASSED test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-True] Stderr: Container roottestconcurrentthreadssoftlimit-gw2-node2-1 Stopping Stderr: Container roottestconcurrentthreadssoftlimit-gw2-node4-1 Stopping Stderr: Container roottestconcurrentthreadssoftlimit-gw2-node3-1 Stopping Stderr: Container roottestconcurrentthreadssoftlimit-gw2-node1-1 Stopping Stderr: Container roottestconcurrentthreadssoftlimit-gw2-node3-1 Stopped Stderr: Container roottestconcurrentthreadssoftlimit-gw2-node4-1 Stopped Stderr: Container roottestconcurrentthreadssoftlimit-gw2-node1-1 Stopped Stderr: Container roottestconcurrentthreadssoftlimit-gw2-node2-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_concurrent_threads_soft_limit/_instances-0-gw2/node1/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_concurrent_threads_soft_limit/_instances-0-gw2/node1/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_concurrent_threads_soft_limit/_instances-0-gw2/node2/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_concurrent_threads_soft_limit/_instances-0-gw2/node2/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_concurrent_threads_soft_limit/_instances-0-gw2/node3/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_concurrent_threads_soft_limit/_instances-0-gw2/node3/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_concurrent_threads_soft_limit/_instances-0-gw2/node4/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_concurrent_threads_soft_limit/_instances-0-gw2/node4/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_concurrent_threads_soft_limit/_instances-0-gw2/.env --project-name roottestconcurrentthreadssoftlimit-gw2 --file /ClickHouse/tests/integration/test_concurrent_threads_soft_limit/_instances-0-gw2/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_concurrent_threads_soft_limit/_instances-0-gw2/node2/docker-compose.yml --file /ClickHouse/tests/integration/test_concurrent_threads_soft_limit/_instances-0-gw2/node3/docker-compose.yml --file /ClickHouse/tests/integration/test_concurrent_threads_soft_limit/_instances-0-gw2/node4/docker-compose.yml down --volumes] http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None Executing query SYSTEM FLUSH LOGS on node1 run container_id:roottestconfigsubstitutions-gw3-node3-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestconfigsubstitutions-gw3-node3-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None run container_id:roottestconfigsubstitutions-gw3-node3-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestconfigsubstitutions-gw3-node3-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query SELECT COUNT() FROM non_replicated_mt on node1 No clickhouse process running. Start new one. http://localhost:None "POST /v1.46/containers/roottestconfigsubstitutions-gw3-node3-1/exec HTTP/1.1" 201 74 http://localhost:None "GET /v1.46/containers/2454c466213b1bea59ee5ab00a8ade0706d2c829ef7687c79b34b2c66494061b/json HTTP/1.1" 200 None ClickHouse node1 started get_instance_ip instance_name=node2 http://localhost:None "GET /v1.46/containers/roottestclusterallreplicas-gw1-node2-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node2 http://localhost:None "GET /v1.46/containers/roottestclusterallreplicas-gw1-node2-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node2, ip: 172.16.8.3... http://localhost:None "GET /v1.46/containers/roottestclusterallreplicas-gw1-node2-1/json HTTP/1.1" 200 None http://localhost:None "POST /v1.46/exec/31f14b30aa3ebb480388c53aaf7cc9c46c5de0041f3957014e8bf3a32e6f134a/start HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/containers/cdeec82a6e984eeb8436ad275e0ac2fd7151452ea77fdf0dd916d9253e20fe5b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/exec/31f14b30aa3ebb480388c53aaf7cc9c46c5de0041f3957014e8bf3a32e6f134a/json HTTP/1.1" 200 586 ClickHouse node2 started Executing query SELECT hostName() FROM clusterAllReplicas('one_shard_two_nodes', system.one) ORDER BY ALL on node1 Executing query SELECT path FROM system.parts WHERE database = 'default' AND table = 'non_replicated_mt' AND name = '201902_1_1_0' on node1 Executing query SELECT max(value) FROM ( SELECT toStartOfInterval(event_time, toIntervalSecond(1)) AS t, avg(value) AS value FROM system.asynchronous_metric_log WHERE event_time >= now() - 60 AND metric = 'OSUserTimeNormalized' GROUP BY t ) SETTINGS max_threads = 1 on node1 Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 run container_id:roottestchecktable-gw4-node1-1 detach:False nothrow:False cmd: ['bash', '-c', 'rm -r /var/lib/clickhouse/store/2dd/2dd39276-fe43-4b78-b3c8-f0e157db2102/201902_1_1_0//checksums.txt'] Command:[docker exec --privileged roottestchecktable-gw4-node1-1 bash -c rm -r /var/lib/clickhouse/store/2dd/2dd39276-fe43-4b78-b3c8-f0e157db2102/201902_1_1_0//checksums.txt] Executing query SELECT hostName(), * FROM clusterAllReplicas("one_shard_two_nodes", system.one) ORDER BY ALL on node1 Stderr: Container roottestallowedclienthosts-gw8-clientA1-1 Stopping Stderr: Container roottestallowedclienthosts-gw8-clientA2-1 Stopping Stderr: Container roottestallowedclienthosts-gw8-clientA3-1 Stopping Stderr: Container roottestallowedclienthosts-gw8-server-1 Stopping Stderr: Container roottestallowedclienthosts-gw8-server-1 Stopped Stderr: Container roottestallowedclienthosts-gw8-clientA3-1 Stopped Stderr: Container roottestallowedclienthosts-gw8-clientA2-1 Stopped Stderr: Container roottestallowedclienthosts-gw8-clientA1-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_allowed_client_hosts/_instances-0-gw8/server/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_allowed_client_hosts/_instances-0-gw8/server/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Executing query SYSTEM FLUSH LOGS on node2 Command:[bash -c [ -f /ClickHouse/tests/integration/test_allowed_client_hosts/_instances-0-gw8/clientA1/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_allowed_client_hosts/_instances-0-gw8/clientA1/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_allowed_client_hosts/_instances-0-gw8/clientA2/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_allowed_client_hosts/_instances-0-gw8/clientA2/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Executing query CHECK TABLE non_replicated_mt PARTITION 201902 on node1 Command:[bash -c [ -f /ClickHouse/tests/integration/test_allowed_client_hosts/_instances-0-gw8/clientA3/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_allowed_client_hosts/_instances-0-gw8/clientA3/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_allowed_client_hosts/_instances-0-gw8/.env --project-name roottestallowedclienthosts-gw8 --file /ClickHouse/tests/integration/test_allowed_client_hosts/_instances-0-gw8/server/docker-compose.yml --file /ClickHouse/tests/integration/test_allowed_client_hosts/_instances-0-gw8/clientA1/docker-compose.yml --file /ClickHouse/tests/integration/test_allowed_client_hosts/_instances-0-gw8/clientA2/docker-compose.yml --file /ClickHouse/tests/integration/test_allowed_client_hosts/_instances-0-gw8/clientA3/docker-compose.yml down --volumes] Executing query SELECT hostName() FROM cluster('one_shard_two_nodes', system.one) on node1 Executing query SELECT max(value) FROM ( SELECT toStartOfInterval(event_time, toIntervalSecond(1)) AS t, avg(value) AS value FROM system.asynchronous_metric_log WHERE event_time >= now() - 60 AND metric = 'OSUserTimeNormalized' GROUP BY t ) SETTINGS max_threads = 1 on node2 Executing query SELECT hostName() FROM cluster('one_shard_two_nodes', system.one) on node2 Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 [gw6] PASSED test_async_metrics_in_cgroup/test.py::test_normalized_user_cpu test_async_metrics_in_cgroup/test.py::test_system_wide_metrics Executing query SELECT value FROM system.build_options WHERE name = 'CXX_FLAGS' on node1 [gw1] PASSED test_cluster_all_replicas/test.py::test_cluster test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[one_shard_three_nodes] Executing query SELECT hostName() FROM clusterAllReplicas('one_shard_three_nodes', system.one) ORDER BY ALL settings skip_unavailable_shards=0 on node1 Stderr: Container roottestconcurrentthreadssoftlimit-gw2-node3-1 Stopping Stderr: Container roottestconcurrentthreadssoftlimit-gw2-node4-1 Stopping Stderr: Container roottestconcurrentthreadssoftlimit-gw2-node1-1 Stopping Stderr: Container roottestconcurrentthreadssoftlimit-gw2-node2-1 Stopping Stderr: Container roottestconcurrentthreadssoftlimit-gw2-node3-1 Stopped Stderr: Container roottestconcurrentthreadssoftlimit-gw2-node3-1 Removing Stderr: Container roottestconcurrentthreadssoftlimit-gw2-node4-1 Stopped Stderr: Container roottestconcurrentthreadssoftlimit-gw2-node4-1 Removing Stderr: Container roottestconcurrentthreadssoftlimit-gw2-node1-1 Stopped Stderr: Container roottestconcurrentthreadssoftlimit-gw2-node1-1 Removing Stderr: Container roottestconcurrentthreadssoftlimit-gw2-node2-1 Stopped Stderr: Container roottestconcurrentthreadssoftlimit-gw2-node2-1 Removing Stderr: Container roottestconcurrentthreadssoftlimit-gw2-node3-1 Removed Stderr: Container roottestconcurrentthreadssoftlimit-gw2-node4-1 Removed Stderr: Container roottestconcurrentthreadssoftlimit-gw2-node2-1 Removed Stderr: Container roottestconcurrentthreadssoftlimit-gw2-node1-1 Removed Stderr: Network roottestconcurrentthreadssoftlimit-gw2_default Removing Stderr: Network roottestconcurrentthreadssoftlimit-gw2_default Removed Cleanup called Executing query SELECT sum(*) FROM system.numbers_mt FORMAT Null SETTINGS max_execution_time=10 on node2 Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Docker networks for project roottestconcurrentthreadssoftlimit-gw2 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestconcurrentthreadssoftlimit-gw2 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestconcurrentthreadssoftlimit-gw2 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestconcurrentthreadssoftlimit-gw2-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestconcurrentthreadssoftlimit-gw2 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:1 Volumes pruned: 1 test_authentication/test.py::test_authentication_fail Running tests in /ClickHouse/tests/integration/test_authentication/test.py Cluster start called. is_up=False Docker networks for project roottestauthentication-gw2 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestauthentication-gw2 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Docker volumes for project roottestauthentication-gw2 are DRIVER VOLUME NAME Cleanup called Docker networks for project roottestauthentication-gw2 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestauthentication-gw2 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestauthentication-gw2 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestauthentication-gw2-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestauthentication-gw2 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:1 Volumes pruned: 1 Setup directory for instance: instance Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files [] to /ClickHouse/tests/integration/test_authentication/_instances-0-gw2/instance/configs/config.d Setup database dir /ClickHouse/tests/integration/test_authentication/_instances-0-gw2/instance/database Setup logs dir /ClickHouse/tests/integration/test_authentication/_instances-0-gw2/instance/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw'} stored in /ClickHouse/tests/integration/test_authentication/_instances-0-gw2/.env Executing query CREATE TABLE tbl ON CLUSTER 'cluster' (x UInt8) ENGINE=ReplicatedMergeTree('/clickhouse/tables/tbl/', '{replica}')ORDER BY x on node1 Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[native-True] No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_authentication/_instances-0-gw2/.env --project-name roottestauthentication-gw2 --file /ClickHouse/tests/integration/test_authentication/_instances-0-gw2/instance/docker-compose.yml pull] Executing query INSERT INTO tbl VALUES (1) on node1 Stderr: Container roottestallowedclienthosts-gw8-clientA2-1 Stopping Stderr: Container roottestallowedclienthosts-gw8-clientA3-1 Stopping Stderr: Container roottestallowedclienthosts-gw8-server-1 Stopping Stderr: Container roottestallowedclienthosts-gw8-clientA1-1 Stopping Stderr: Container roottestallowedclienthosts-gw8-clientA2-1 Stopped Stderr: Container roottestallowedclienthosts-gw8-clientA2-1 Removing Stderr: Container roottestallowedclienthosts-gw8-clientA3-1 Stopped Stderr: Container roottestallowedclienthosts-gw8-clientA3-1 Removing Stderr: Container roottestallowedclienthosts-gw8-server-1 Stopped Stderr: Container roottestallowedclienthosts-gw8-server-1 Removing Stderr: Container roottestallowedclienthosts-gw8-clientA1-1 Stopped Stderr: Container roottestallowedclienthosts-gw8-clientA1-1 Removing Stderr: Container roottestallowedclienthosts-gw8-server-1 Removed Stderr: Container roottestallowedclienthosts-gw8-clientA1-1 Removed Stderr: Container roottestallowedclienthosts-gw8-clientA3-1 Removed Stderr: Container roottestallowedclienthosts-gw8-clientA2-1 Removed Stderr: Network roottestallowedclienthosts-gw8_default Removing Stderr: Network roottestallowedclienthosts-gw8_default Removed Cleanup called Docker networks for project roottestallowedclienthosts-gw8 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestallowedclienthosts-gw8 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Executing query BACKUP TABLE tbl ON CLUSTER 'cluster' TO Disk('backups', '3') ASYNC on node1 Docker volumes for project roottestallowedclienthosts-gw8 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestallowedclienthosts-gw8-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestallowedclienthosts-gw8 Trying to prune unused networks... Executing query BACKUP TABLE tbl ON CLUSTER 'cluster' TO Disk('backups', '3') ASYNC on node2 Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] run container_id:roottestconfigsubstitutions-gw3-node3-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestconfigsubstitutions-gw3-node3-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:1 Volumes pruned: 1 Running tests in /ClickHouse/tests/integration/test_attach_partition_with_large_destination/test.py Cluster start called. is_up=False test_attach_partition_with_large_destination/test.py::test_attach_partition_with_large_destination[Atomic] Executing query SELECT status FROM system.backups WHERE id IN ['af7d86d8-1699-4ec4-8f4e-0535277eb9f3', '04fc6848-57ec-4ca8-8187-dc08ed798cc1'] AND status == 'CREATING_BACKUP' on node1 Docker networks for project roottestattachpartitionwithlargedestination-gw8 are NETWORK ID NAME DRIVER SCOPE Stdout:776 Clickhouse process running. run container_id:roottestconfigsubstitutions-gw3-node3-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestconfigsubstitutions-gw3-node3-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Docker containers for project roottestattachpartitionwithlargedestination-gw8 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestattachpartitionwithlargedestination-gw8 are DRIVER VOLUME NAME Cleanup called Docker networks for project roottestattachpartitionwithlargedestination-gw8 are NETWORK ID NAME DRIVER SCOPE Stdout:776 Executing query select 20 on node3 Docker containers for project roottestattachpartitionwithlargedestination-gw8 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestattachpartitionwithlargedestination-gw8 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestattachpartitionwithlargedestination-gw8-.*-1$' --format '{{.ID}}:{{.Names}}'] Executing query SELECT status FROM system.backups WHERE id IN ['af7d86d8-1699-4ec4-8f4e-0535277eb9f3', '04fc6848-57ec-4ca8-8187-dc08ed798cc1'] AND status == 'CREATING_BACKUP' on node2 Executing query SYSTEM RELOAD CONFIG on node3 Unstopped containers: {} No running containers for project: roottestattachpartitionwithlargedestination-gw8 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Executing query SELECT count() FROM system.backups WHERE id IN ['af7d86d8-1699-4ec4-8f4e-0535277eb9f3', '04fc6848-57ec-4ca8-8187-dc08ed798cc1'] AND status == 'BACKUP_CREATED' on node1 Executing query SELECT * FROM test.tbl3 ORDER BY p on node Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Executing query SELECT COUNT() FROM non_replicated_mt on node1 Executing query SELECT value FROM system.merge_tree_settings WHERE name='min_bytes_for_wide_part' on node3 Stdout:1 Volumes pruned: 1 Setup directory for instance: node Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_attach_partition_with_large_destination/configs/config.xml'] to /ClickHouse/tests/integration/test_attach_partition_with_large_destination/_instances-0-gw8/node/configs/config.d Setup database dir /ClickHouse/tests/integration/test_attach_partition_with_large_destination/_instances-0-gw8/node/database Setup logs dir /ClickHouse/tests/integration/test_attach_partition_with_large_destination/_instances-0-gw8/node/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'keeper_binary': '/clickhouse', 'keeper_cmd_prefix': 'clickhouse keeper', 'image': 'altinityinfra/integration-test:8b2301119731', 'user': '0', 'keeper_fs': 'bind', 'keeper_logs_dir1': '/ClickHouse/tests/integration/test_attach_partition_with_large_destination/_instances-0-gw8/keeper1/log', 'keeper_config_dir1': '/ClickHouse/tests/integration/test_attach_partition_with_large_destination/_instances-0-gw8/keeper1/config', 'keeper_db_dir1': '/ClickHouse/tests/integration/test_attach_partition_with_large_destination/_instances-0-gw8/keeper1/coordination', 'keeper_logs_dir2': '/ClickHouse/tests/integration/test_attach_partition_with_large_destination/_instances-0-gw8/keeper2/log', 'keeper_config_dir2': '/ClickHouse/tests/integration/test_attach_partition_with_large_destination/_instances-0-gw8/keeper2/config', 'keeper_db_dir2': '/ClickHouse/tests/integration/test_attach_partition_with_large_destination/_instances-0-gw8/keeper2/coordination', 'keeper_logs_dir3': '/ClickHouse/tests/integration/test_attach_partition_with_large_destination/_instances-0-gw8/keeper3/log', 'keeper_config_dir3': '/ClickHouse/tests/integration/test_attach_partition_with_large_destination/_instances-0-gw8/keeper3/config', 'keeper_db_dir3': '/ClickHouse/tests/integration/test_attach_partition_with_large_destination/_instances-0-gw8/keeper3/coordination'} stored in /ClickHouse/tests/integration/test_attach_partition_with_large_destination/_instances-0-gw8/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_attach_partition_with_large_destination/_instances-0-gw8/.env --project-name roottestattachpartitionwithlargedestination-gw8 --file /ClickHouse/tests/integration/test_attach_partition_with_large_destination/_instances-0-gw8/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml pull] Executing query SELECT count() FROM system.backups WHERE id IN ['af7d86d8-1699-4ec4-8f4e-0535277eb9f3', '04fc6848-57ec-4ca8-8187-dc08ed798cc1'] AND status == 'BACKUP_CREATED' on node2 Executing query DROP TABLE IF EXISTS test.tbl3 on node Executing query SELECT value FROM system.merge_tree_settings WHERE name='min_rows_for_wide_part' on node3 Executing query SELECT count() FROM system.backups WHERE id IN ['af7d86d8-1699-4ec4-8f4e-0535277eb9f3', '04fc6848-57ec-4ca8-8187-dc08ed798cc1'] AND status == 'BACKUP_FAILED' on node1 Executing query CREATE TABLE test.tbl1 AS test.tbl on node [gw5] PASSED test_backup_restore/test.py::test_replace_partition test_backup_restore/test.py::test_restore Executing query SELECT path FROM system.parts WHERE table = 'non_replicated_mt' and name = '201902_1_1_0' AND database = 'default' on node1 Executing query SELECT value FROM system.merge_tree_settings WHERE name='merge_max_block_size' on node3 copy from /var/lib/clickhouse/shadow/*/data/test/tbl to /var/lib/clickhouse/data/test/tbl1/detached run container_id:roottestbackuprestore-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', 'cp -r /var/lib/clickhouse/shadow/*/data/test/tbl -T /var/lib/clickhouse/data/test/tbl1/detached'] Command:[docker exec roottestbackuprestore-gw5-node-1 bash -c cp -r /var/lib/clickhouse/shadow/*/data/test/tbl -T /var/lib/clickhouse/data/test/tbl1/detached] Executing query SELECT value FROM system.server_settings WHERE name='background_pool_size' on node3 Executing query SELECT count() FROM system.backups WHERE id IN ['af7d86d8-1699-4ec4-8f4e-0535277eb9f3', '04fc6848-57ec-4ca8-8187-dc08ed798cc1'] AND status == 'BACKUP_FAILED' on node2 Executing query ALTER TABLE test.tbl1 ATTACH PARTITION 197001 on node run container_id:roottestchecktable-gw4-node1-1 detach:False nothrow:False cmd: ['bash', '-c', 'cd /var/lib/clickhouse/store/2dd/2dd39276-fe43-4b78-b3c8-f0e157db2102/201902_1_1_0/ && ls *.bin | head -n 1'] Command:[docker exec roottestchecktable-gw4-node1-1 bash -c cd /var/lib/clickhouse/store/2dd/2dd39276-fe43-4b78-b3c8-f0e157db2102/201902_1_1_0/ && ls *.bin | head -n 1] get_instance_ip instance_name=zoo1 http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw3-zoo1-1/json HTTP/1.1" 200 None get_kazoo_client: zoo1, ip:172.16.2.3, port:2181, use_ssl:False Connecting to 172.16.2.3(172.16.2.3):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Sending request(xid=1): Create(path='/background_pool_size', data=b'72', acl=[ACL(perms=31, acl_list=['ALL'], id=Id(scheme='world', id='anyone'))], flags=0) Received response(xid=1): '/background_pool_size' run container_id:roottestconfigsubstitutions-gw3-node3-1 detach:False nothrow:False cmd: ['bash', '-c', 'test ! -f /etc/clickhouse-server/config.d/config_zk_include_test.xml || mv --no-clobber /etc/clickhouse-server/config.d/config_zk_include_test.xml /tmp/config_zk_include_test_34ca4601-9d6d-427b-b4d9-ef0f9f823e9c.xml'] Command:[docker exec roottestconfigsubstitutions-gw3-node3-1 bash -c test ! -f /etc/clickhouse-server/config.d/config_zk_include_test.xml || mv --no-clobber /etc/clickhouse-server/config.d/config_zk_include_test.xml /tmp/config_zk_include_test_34ca4601-9d6d-427b-b4d9-ef0f9f823e9c.xml] Executing query DROP TABLE tbl ON CLUSTER 'cluster' SYNC on node1 Stdout:data.bin run container_id:roottestchecktable-gw4-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "cd /var/lib/clickhouse/store/2dd/2dd39276-fe43-4b78-b3c8-f0e157db2102/201902_1_1_0/ && ls *.bin | head -n 1 | xargs -I{} sh -c 'truncate -s -1 $1' -- {}"] Command:[docker exec --privileged roottestchecktable-gw4-node1-1 bash -c cd /var/lib/clickhouse/store/2dd/2dd39276-fe43-4b78-b3c8-f0e157db2102/201902_1_1_0/ && ls *.bin | head -n 1 | xargs -I{} sh -c 'truncate -s -1 $1' -- {}] Executing query ALTER TABLE test.tbl1 ATTACH PARTITION 197002 on node run container_id:roottestconfigsubstitutions-gw3-node3-1 detach:False nothrow:False cmd: ['bash', '-c', 'echo \'\n \n \n 44\n \n \n 1\n 1111\n \n\n \n \n \' > /etc/clickhouse-server/config.d/config_zk_include_test.xml'] Command:[docker exec roottestconfigsubstitutions-gw3-node3-1 bash -c echo ' 44 1 1111 ' > /etc/clickhouse-server/config.d/config_zk_include_test.xml] Executing query CHECK TABLE non_replicated_mt on node1 Executing query SELECT sleep(2) on node Executing query SYSTEM RELOAD CONFIG on node3 Executing query RESTORE TABLE tbl FROM Disk('backups', '3') on node1 Executing query SELECT value FROM system.server_settings WHERE name='background_pool_size' on node3 Executing query SELECT * FROM tbl on node1 run container_id:roottestconfigsubstitutions-gw3-node3-1 detach:False nothrow:False cmd: ['bash', '-c', 'test ! -f /tmp/config_zk_include_test_34ca4601-9d6d-427b-b4d9-ef0f9f823e9c.xml || mv /tmp/config_zk_include_test_34ca4601-9d6d-427b-b4d9-ef0f9f823e9c.xml /etc/clickhouse-server/config.d/config_zk_include_test.xml'] Command:[docker exec roottestconfigsubstitutions-gw3-node3-1 bash -c test ! -f /tmp/config_zk_include_test_34ca4601-9d6d-427b-b4d9-ef0f9f823e9c.xml || mv /tmp/config_zk_include_test_34ca4601-9d6d-427b-b4d9-ef0f9f823e9c.xml /etc/clickhouse-server/config.d/config_zk_include_test.xml] Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw0] PASSED test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[native-True] Sending request(xid=2): Delete(path='/background_pool_size', version=-1) Received response(xid=2): True Executing query select 1 on node4 [gw3] PASSED test_config_substitutions/test.py::test_config_multiple_zk_substitutions test_config_substitutions/test.py::test_include_config Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select 1 on node4 Executing query select 1 on node4 Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Executing query select 1 on node3 Executing query select 1 on node3 Executing query select 1 on node3 Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select 1 on node8 Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Executing query select 1 on node8 Executing query select 1 on node8 test_backup_restore_on_cluster/test.py::test_backup_restore_on_single_replica Executing query CREATE DATABASE mydb ON CLUSTER 'cluster' ENGINE=Replicated('/clickhouse/path/','{shard}','{replica}') on node1 Command:[docker compose --env-file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/.env --project-name roottestconfigsubstitutions-gw3 --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node2/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node3/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node4/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node5/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node6/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node7/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node8/docker-compose.yml stop --timeout 20] [gw3] PASSED test_config_substitutions/test.py::test_include_config Executing query CREATE TABLE mydb.test (`name` String, `value` UInt32) ENGINE = ReplicatedMergeTree ORDER BY value on node1 Executing query INSERT INTO mydb.test VALUES ('abc', 1), ('def', 2) on node1 Executing query INSERT INTO mydb.test VALUES ('ghi', 3) on node1 Executing query BACKUP DATABASE mydb TO Disk('backups', '4') on node1 Executing query DROP DATABASE mydb SYNC on node1 Executing query CHECK TABLE non_replicated_mt on node1 Executing query RESTORE DATABASE mydb FROM Disk('backups', '4') on node1 Executing query RESTORE DATABASE mydb FROM Disk('backups', '4') SETTINGS structure_only=true on node1 Executing query SYSTEM SYNC REPLICA mydb.test on node1 Executing query SELECT * FROM mydb.test ORDER BY name on node1 Executing query DROP DATABASE mydb SYNC on node1 Executing query RESTORE DATABASE mydb FROM Disk('backups', '4') SETTINGS allow_non_empty_tables=true on node1 Executing query SELECT * FROM test.tbl1 ORDER BY p on node Executing query SYSTEM SYNC REPLICA mydb.test on node1 Executing query ALTER TABLE test.tbl1 UPDATE k=10 WHERE 1 on node Executing query SELECT * FROM mydb.test ORDER BY name on node1 Executing query SELECT sleep(2) on node Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw0] PASSED test_backup_restore_on_cluster/test.py::test_backup_restore_on_single_replica Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Executing query INSERT INTO non_replicated_mt VALUES (toDate('2019-01-01'), 1, 10), (toDate('2019-01-01'), 2, 12) on node1 Executing query CHECK TABLE non_replicated_mt PARTITION 201901 on node1 Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Executing query SELECT path FROM system.parts WHERE table = 'non_replicated_mt' and name = '201901_2_2_0' AND database = 'default' on node1 Executing query CREATE TABLE tbl (`x` UInt8, `y` String) ENGINE = MergeTree ORDER BY x on node1 run container_id:roottestchecktable-gw4-node1-1 detach:False nothrow:False cmd: ['bash', '-c', 'cd /var/lib/clickhouse/store/2dd/2dd39276-fe43-4b78-b3c8-f0e157db2102/201901_2_2_0/ && ls *.bin | head -n 1'] Command:[docker exec roottestchecktable-gw4-node1-1 bash -c cd /var/lib/clickhouse/store/2dd/2dd39276-fe43-4b78-b3c8-f0e157db2102/201901_2_2_0/ && ls *.bin | head -n 1] test_backup_restore_on_cluster/test.py::test_different_tables_on_nodes Executing query CREATE TABLE tbl (`w` Int64) ENGINE = MergeTree ORDER BY w on node2 Stdout:data.bin run container_id:roottestchecktable-gw4-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "cd /var/lib/clickhouse/store/2dd/2dd39276-fe43-4b78-b3c8-f0e157db2102/201901_2_2_0/ && ls *.bin | head -n 1 | xargs -I{} sh -c 'truncate -s -1 $1' -- {}"] Command:[docker exec --privileged roottestchecktable-gw4-node1-1 bash -c cd /var/lib/clickhouse/store/2dd/2dd39276-fe43-4b78-b3c8-f0e157db2102/201901_2_2_0/ && ls *.bin | head -n 1 | xargs -I{} sh -c 'truncate -s -1 $1' -- {}] Executing query INSERT INTO tbl VALUES (1, 'Don''t'), (2, 'count'), (3, 'your'), (4, 'chickens') on node1 Executing query SELECT path FROM system.parts WHERE database = 'default' AND table = 'non_replicated_mt' AND name = '201901_2_2_0' on node1 Executing query INSERT INTO tbl VALUES (-333), (-222), (-111), (0), (111) on node2 Executing query BACKUP TABLE tbl ON CLUSTER 'cluster' TO Disk('backups', '5') on node1 run container_id:roottestchecktable-gw4-node1-1 detach:False nothrow:False cmd: ['bash', '-c', 'rm -r /var/lib/clickhouse/store/2dd/2dd39276-fe43-4b78-b3c8-f0e157db2102/201901_2_2_0//checksums.txt'] Command:[docker exec --privileged roottestchecktable-gw4-node1-1 bash -c rm -r /var/lib/clickhouse/store/2dd/2dd39276-fe43-4b78-b3c8-f0e157db2102/201901_2_2_0//checksums.txt] Executing query CHECK TABLE non_replicated_mt PARTITION 201901 on node1 Executing query DROP TABLE tbl ON CLUSTER 'cluster' SYNC on node1 Executing query RESTORE TABLE tbl ON CLUSTER 'cluster' FROM Disk('backups', '5') on node2 Executing query SELECT * FROM tbl on node1 Executing query SELECT * FROM tbl on node2 Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw0] PASSED test_backup_restore_on_cluster/test.py::test_different_tables_on_nodes Executing query SELECT * FROM test.tbl1 ORDER BY p on node Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Executing query DROP TABLE IF EXISTS test.tbl1 on node Command:[docker compose --env-file /ClickHouse/tests/integration/test_backup_restore/_instances-0-gw5/.env --project-name roottestbackuprestore-gw5 --file /ClickHouse/tests/integration/test_backup_restore/_instances-0-gw5/node/docker-compose.yml stop --timeout 20] [gw5] PASSED test_backup_restore/test.py::test_restore Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 test_backup_restore_on_cluster/test.py::test_empty_replicated_table Executing query CREATE TABLE tbl ON CLUSTER 'cluster' (x UInt8, y String) ENGINE=ReplicatedMergeTree('/clickhouse/tables/tbl/', '{replica}')ORDER BY x on node1 Executing query BACKUP TABLE tbl ON CLUSTER 'cluster' TO Disk('backups', '6') SETTINGS replica_num=1 on node1 Executing query DROP TABLE tbl ON CLUSTER 'cluster' SYNC on node1 Stderr: instance Pulling Stderr: instance Pulled ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_authentication/_instances-0-gw2/.env --project-name roottestauthentication-gw2 --file /ClickHouse/tests/integration/test_authentication/_instances-0-gw2/instance/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_authentication/_instances-0-gw2/.env --project-name roottestauthentication-gw2 --file /ClickHouse/tests/integration/test_authentication/_instances-0-gw2/instance/docker-compose.yml up -d --no-recreate] Stderr: node Pulling Stderr: node Pulled ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_attach_table_normalizer/_instances-0-gw9/.env --project-name roottestattachtablenormalizer-gw9 --file /ClickHouse/tests/integration/test_attach_table_normalizer/_instances-0-gw9/node/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_attach_table_normalizer/_instances-0-gw9/.env --project-name roottestattachtablenormalizer-gw9 --file /ClickHouse/tests/integration/test_attach_table_normalizer/_instances-0-gw9/node/docker-compose.yml up -d --no-recreate] Stderr: zoo3 Skipped - Image is already being pulled by zoo2 Stderr: node2 Skipped - Image is already being pulled by zoo2 Stderr: node3 Skipped - Image is already being pulled by zoo2 Stderr: node4 Skipped - Image is already being pulled by zoo2 Stderr: node1 Skipped - Image is already being pulled by zoo2 Stderr: zoo1 Skipped - Image is already being pulled by zoo2 Stderr: zoo2 Pulling Stderr: zoo2 Pulled Setup ZooKeeper Creating internal ZooKeeper dirs: ['/ClickHouse/tests/integration/test_alter_on_mixed_type_cluster/_instances-0-gw7/keeper1/log', '/ClickHouse/tests/integration/test_alter_on_mixed_type_cluster/_instances-0-gw7/keeper1/config', '/ClickHouse/tests/integration/test_alter_on_mixed_type_cluster/_instances-0-gw7/keeper1/coordination', '/ClickHouse/tests/integration/test_alter_on_mixed_type_cluster/_instances-0-gw7/keeper2/log', '/ClickHouse/tests/integration/test_alter_on_mixed_type_cluster/_instances-0-gw7/keeper2/config', '/ClickHouse/tests/integration/test_alter_on_mixed_type_cluster/_instances-0-gw7/keeper2/coordination', '/ClickHouse/tests/integration/test_alter_on_mixed_type_cluster/_instances-0-gw7/keeper3/log', '/ClickHouse/tests/integration/test_alter_on_mixed_type_cluster/_instances-0-gw7/keeper3/config', '/ClickHouse/tests/integration/test_alter_on_mixed_type_cluster/_instances-0-gw7/keeper3/coordination'] Command:[docker compose --project-name roottestalteronmixedtypecluster-gw7 --env-file /ClickHouse/tests/integration/test_alter_on_mixed_type_cluster/_instances-0-gw7/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --verbose up -d] [gw4] PASSED test_check_table/test.py::test_check_normal_table_corruption[] test_check_table/test.py::test_check_replicated_table_corruption[-_0-.bin] Executing query DROP TABLE IF EXISTS replicated_mt_1 SYNC on node1 Executing query RESTORE TABLE tbl ON CLUSTER 'cluster' FROM Disk('backups', '6') on node1 Executing query CREATE TABLE replicated_mt_1(date Date, id UInt32, value Int32) ENGINE = ReplicatedMergeTree('/clickhouse/tables/replicated_mt_1__0', 'node1') PARTITION BY toYYYYMM(date) ORDER BY id on node1 Stderr: zoo1 Skipped - Image is already being pulled by zoo3 Stderr: zoo2 Skipped - Image is already being pulled by zoo3 Stderr: node Skipped - Image is already being pulled by zoo3 Stderr: zoo3 Pulling Stderr: zoo3 Pulled Setup ZooKeeper Creating internal ZooKeeper dirs: ['/ClickHouse/tests/integration/test_attach_partition_with_large_destination/_instances-0-gw8/keeper1/log', '/ClickHouse/tests/integration/test_attach_partition_with_large_destination/_instances-0-gw8/keeper1/config', '/ClickHouse/tests/integration/test_attach_partition_with_large_destination/_instances-0-gw8/keeper1/coordination', '/ClickHouse/tests/integration/test_attach_partition_with_large_destination/_instances-0-gw8/keeper2/log', '/ClickHouse/tests/integration/test_attach_partition_with_large_destination/_instances-0-gw8/keeper2/config', '/ClickHouse/tests/integration/test_attach_partition_with_large_destination/_instances-0-gw8/keeper2/coordination', '/ClickHouse/tests/integration/test_attach_partition_with_large_destination/_instances-0-gw8/keeper3/log', '/ClickHouse/tests/integration/test_attach_partition_with_large_destination/_instances-0-gw8/keeper3/config', '/ClickHouse/tests/integration/test_attach_partition_with_large_destination/_instances-0-gw8/keeper3/coordination'] Command:[docker compose --project-name roottestattachpartitionwithlargedestination-gw8 --env-file /ClickHouse/tests/integration/test_attach_partition_with_large_destination/_instances-0-gw8/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --verbose up -d] Executing query DROP TABLE IF EXISTS replicated_mt_1 SYNC on node2 Executing query CREATE TABLE replicated_mt_1(date Date, id UInt32, value Int32) ENGINE = ReplicatedMergeTree('/clickhouse/tables/replicated_mt_1__0', 'node2') PARTITION BY toYYYYMM(date) ORDER BY id on node2 Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster' tbl on node1 Executing query INSERT INTO replicated_mt_1 VALUES (toDate('2019-02-01'), 1, 10), (toDate('2019-02-01'), 2, 12) on node1 Executing query INSERT INTO replicated_mt_1 VALUES (toDate('2019-01-02'), 3, 10), (toDate('2019-01-02'), 4, 12) on node1 Executing query SELECT * FROM tbl on node1 Executing query SYSTEM SYNC REPLICA replicated_mt_1 on node2 Executing query SELECT count() from replicated_mt_1 on node1 Executing query SELECT * FROM tbl on node2 Stderr: Network roottestauthentication-gw2_default Creating Stderr: Network roottestauthentication-gw2_default Created Stderr: Container roottestauthentication-gw2-instance-1 Creating Stderr: Container roottestauthentication-gw2-instance-1 Created Stderr: Container roottestauthentication-gw2-instance-1 Starting Stderr: Container roottestauthentication-gw2-instance-1 Started ClickHouse instance created get_instance_ip instance_name=instance http://localhost:None "GET /v1.46/containers/roottestauthentication-gw2-instance-1/json HTTP/1.1" 200 None Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 get_instance_ip instance_name=instance http://localhost:None "GET /v1.46/containers/roottestauthentication-gw2-instance-1/json HTTP/1.1" 200 None [gw0] PASSED test_backup_restore_on_cluster/test.py::test_empty_replicated_table Waiting for ClickHouse start in instance, ip: 172.16.1.2... Executing query SELECT count() from replicated_mt_1 on node2 http://localhost:None "GET /v1.46/containers/roottestauthentication-gw2-instance-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/e827786440885d628a8cf52f82b60eff3dd1ea98abcb2b31fc2e04cf58ed3a81/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/e827786440885d628a8cf52f82b60eff3dd1ea98abcb2b31fc2e04cf58ed3a81/json HTTP/1.1" 200 None Executing query SELECT name from system.parts where table = 'replicated_mt_1' and partition_id = '201901' and active = 1 on node1 Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 http://localhost:None "GET /v1.46/containers/e827786440885d628a8cf52f82b60eff3dd1ea98abcb2b31fc2e04cf58ed3a81/json HTTP/1.1" 200 None Executing query SELECT path FROM system.parts WHERE table = 'replicated_mt_1' and name = '201901_0_0_0' AND database = 'default' on node1 Stderr: Network roottestattachtablenormalizer-gw9_default Creating Stderr: Network roottestattachtablenormalizer-gw9_default Created Stderr: Container roottestattachtablenormalizer-gw9-node-1 Creating Stderr: Container roottestattachtablenormalizer-gw9-node-1 Created Stderr: Container roottestattachtablenormalizer-gw9-node-1 Starting Stderr: Container roottestattachtablenormalizer-gw9-node-1 Started ClickHouse instance created get_instance_ip instance_name=node http://localhost:None "GET /v1.46/containers/roottestattachtablenormalizer-gw9-node-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node http://localhost:None "GET /v1.46/containers/roottestattachtablenormalizer-gw9-node-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node, ip: 172.16.4.2... http://localhost:None "GET /v1.46/containers/roottestattachtablenormalizer-gw9-node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/9891b4288355de684a42312fadb22ab0ab076ef089672e9c549d5cf36c0829de/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/e827786440885d628a8cf52f82b60eff3dd1ea98abcb2b31fc2e04cf58ed3a81/json HTTP/1.1" 200 None Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 run container_id:roottestchecktable-gw4-node1-1 detach:False nothrow:False cmd: ['bash', '-c', 'cd /var/lib/clickhouse/store/a92/a92cf71c-b4bc-4c65-ba9a-dbc651a411fa/201901_0_0_0/ && ls *.bin | head -n 1'] Command:[docker exec roottestchecktable-gw4-node1-1 bash -c cd /var/lib/clickhouse/store/a92/a92cf71c-b4bc-4c65-ba9a-dbc651a411fa/201901_0_0_0/ && ls *.bin | head -n 1] http://localhost:None "GET /v1.46/containers/9891b4288355de684a42312fadb22ab0ab076ef089672e9c549d5cf36c0829de/json HTTP/1.1" 200 None [gw1] PASSED test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[one_shard_three_nodes] Executing query SELECT hostName() FROM clusterAllReplicas('two_shards_three_nodes', system.one) ORDER BY ALL settings skip_unavailable_shards=0 on node1 test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[two_shards_three_nodes] http://localhost:None "GET /v1.46/containers/e827786440885d628a8cf52f82b60eff3dd1ea98abcb2b31fc2e04cf58ed3a81/json HTTP/1.1" 200 None Stdout:data.bin run container_id:roottestchecktable-gw4-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "cd /var/lib/clickhouse/store/a92/a92cf71c-b4bc-4c65-ba9a-dbc651a411fa/201901_0_0_0/ && ls *.bin | head -n 1 | xargs -I{} sh -c 'truncate -s -1 $1' -- {}"] Command:[docker exec --privileged roottestchecktable-gw4-node1-1 bash -c cd /var/lib/clickhouse/store/a92/a92cf71c-b4bc-4c65-ba9a-dbc651a411fa/201901_0_0_0/ && ls *.bin | head -n 1 | xargs -I{} sh -c 'truncate -s -1 $1' -- {}] Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 http://localhost:None "GET /v1.46/containers/9891b4288355de684a42312fadb22ab0ab076ef089672e9c549d5cf36c0829de/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/e827786440885d628a8cf52f82b60eff3dd1ea98abcb2b31fc2e04cf58ed3a81/json HTTP/1.1" 200 None Executing query CHECK TABLE replicated_mt_1 PARTITION 201901 on node1 http://localhost:None "GET /v1.46/containers/9891b4288355de684a42312fadb22ab0ab076ef089672e9c549d5cf36c0829de/json HTTP/1.1" 200 None Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 http://localhost:None "GET /v1.46/containers/e827786440885d628a8cf52f82b60eff3dd1ea98abcb2b31fc2e04cf58ed3a81/json HTTP/1.1" 200 None ClickHouse instance started Executing query CREATE USER sasha on instance http://localhost:None "GET /v1.46/containers/9891b4288355de684a42312fadb22ab0ab076ef089672e9c549d5cf36c0829de/json HTTP/1.1" 200 None Stderr:time="2025-04-02T03:34:06Z" level=trace msg="Docker Desktop integration not enabled" Stderr: Network roottestalteronmixedtypecluster-gw7_default Creating Stderr: Network roottestalteronmixedtypecluster-gw7_default Created Stderr: Container roottestalteronmixedtypecluster-gw7-zoo2-1 Creating Stderr: Container roottestalteronmixedtypecluster-gw7-zoo3-1 Creating Stderr: Container roottestalteronmixedtypecluster-gw7-zoo1-1 Creating Stderr: Container roottestalteronmixedtypecluster-gw7-zoo3-1 Created Stderr: Container roottestalteronmixedtypecluster-gw7-zoo2-1 Created Stderr: Container roottestalteronmixedtypecluster-gw7-zoo1-1 Created Stderr: Container roottestalteronmixedtypecluster-gw7-zoo3-1 Starting Stderr: Container roottestalteronmixedtypecluster-gw7-zoo1-1 Starting Stderr: Container roottestalteronmixedtypecluster-gw7-zoo2-1 Starting Stderr: Container roottestalteronmixedtypecluster-gw7-zoo3-1 Started Stderr: Container roottestalteronmixedtypecluster-gw7-zoo1-1 Started Stderr: Container roottestalteronmixedtypecluster-gw7-zoo2-1 Started Stderr:time="2025-04-02T03:34:07Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T03:34:07Z" level=debug msg="otel error" error="" Wait ZooKeeper to start get_instance_ip instance_name=zoo1 Executing query CREATE USER masha IDENTIFIED BY 'qwerty' on instance http://localhost:None "GET /v1.46/containers/roottestalteronmixedtypecluster-gw7-zoo1-1/json HTTP/1.1" 200 None get_kazoo_client: zoo1, ip:172.16.7.2, port:2181, use_ssl:False Connecting to 172.16.7.2(172.16.7.2):2181, use_ssl: False Connection dropped: socket connection error: Connection refused test_backup_restore_on_cluster/test.py::test_file_deduplication Executing query CREATE TABLE tbl ON CLUSTER 'cluster' ( yPcrXrEIpP Int32 ) ENGINE=ReplicatedMergeTree('/clickhouse/tables/tbl/', '{replica}') ORDER BY tuple() SETTINGS min_bytes_for_wide_part=0 on node1 Executing query SELECT currentUser() on instance http://localhost:None "GET /v1.46/containers/9891b4288355de684a42312fadb22ab0ab076ef089672e9c549d5cf36c0829de/json HTTP/1.1" 200 None Connecting to 172.16.7.2(172.16.7.2):2181, use_ssl: False Connection dropped: socket connection error: Connection refused http://localhost:None "GET /v1.46/containers/9891b4288355de684a42312fadb22ab0ab076ef089672e9c549d5cf36c0829de/json HTTP/1.1" 200 None ClickHouse node started Executing query DROP TABLE IF EXISTS default.file on node Executing query SELECT currentUser() on instance Executing query CREATE TABLE tbl2 ON CLUSTER 'cluster' ( yPcrXrEIpP Int32 ) ENGINE=ReplicatedMergeTree('/clickhouse/tables/tbl/', '{replica}-2') ORDER BY tuple() SETTINGS min_bytes_for_wide_part=0 on node1 Connecting to 172.16.7.2(172.16.7.2):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Executing query CREATE TABLE default.file(`s` String, `n` UInt8) ENGINE = MergeTree PARTITION BY substring(s, 1, 2) ORDER BY n on node Executing query SELECT currentUser() on instance [gw2] PASSED test_authentication/test.py::test_authentication_fail test_authentication/test.py::test_authentication_pass Stderr:time="2025-04-02T03:34:06Z" level=trace msg="Docker Desktop integration not enabled" Stderr: Network roottestattachpartitionwithlargedestination-gw8_default Creating Stderr: Network roottestattachpartitionwithlargedestination-gw8_default Created Stderr: Container roottestattachpartitionwithlargedestination-gw8-zoo3-1 Creating Stderr: Container roottestattachpartitionwithlargedestination-gw8-zoo1-1 Creating Stderr: Container roottestattachpartitionwithlargedestination-gw8-zoo2-1 Creating Stderr: Container roottestattachpartitionwithlargedestination-gw8-zoo1-1 Created Stderr: Container roottestattachpartitionwithlargedestination-gw8-zoo3-1 Created Stderr: Container roottestattachpartitionwithlargedestination-gw8-zoo2-1 Created Stderr: Container roottestattachpartitionwithlargedestination-gw8-zoo2-1 Starting Stderr: Container roottestattachpartitionwithlargedestination-gw8-zoo3-1 Starting Stderr: Container roottestattachpartitionwithlargedestination-gw8-zoo1-1 Starting Stderr: Container roottestattachpartitionwithlargedestination-gw8-zoo2-1 Started Stderr: Container roottestattachpartitionwithlargedestination-gw8-zoo1-1 Started Stderr: Container roottestattachpartitionwithlargedestination-gw8-zoo3-1 Started Stderr:time="2025-04-02T03:34:07Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T03:34:07Z" level=debug msg="otel error" error="" Wait ZooKeeper to start get_instance_ip instance_name=zoo1 http://localhost:None "GET /v1.46/containers/roottestattachpartitionwithlargedestination-gw8-zoo1-1/json HTTP/1.1" 200 None get_kazoo_client: zoo1, ip:172.16.9.4, port:2181, use_ssl:False Connecting to 172.16.9.4(172.16.9.4):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Executing query DETACH TABLE file on node Executing query INSERT INTO tbl VALUES (3556), (1177), (4004), (4264), (3729), (1438), (2158), (2684), (415), (1917) on node1 Executing query SELECT currentUser() on instance run container_id:roottestattachtablenormalizer-gw9-node-1 detach:False nothrow:False cmd: ['bash', '-c', "sed -i 's/substring/substr/g' /var/lib/clickhouse/metadata/default/file.sql"] Command:[docker exec -u root roottestattachtablenormalizer-gw9-node-1 bash -c sed -i 's/substring/substr/g' /var/lib/clickhouse/metadata/default/file.sql] Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster' tbl on node1 Connecting to 172.16.9.4(172.16.9.4):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Executing query ATTACH TABLE file on node Executing query SYSTEM SYNC REPLICA replicated_mt_1 on node1 Executing query SELECT currentUser() on instance [gw9] PASSED test_attach_table_normalizer/test.py::test_attach_substr Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster' tbl2 on node1 test_attach_table_normalizer/test.py::test_attach_substr_restart Executing query DROP TABLE IF EXISTS default.file on node Executing query SELECT currentUser() on instance Executing query CREATE TABLE default.file(`s` String, `n` UInt8) ENGINE = MergeTree PARTITION BY substring(s, 1, 2) ORDER BY n on node run container_id:roottestattachtablenormalizer-gw9-node-1 detach:False nothrow:False cmd: ['bash', '-c', "sed -i 's/substring/substr/g' /var/lib/clickhouse/metadata/default/file.sql"] Command:[docker exec -u root roottestattachtablenormalizer-gw9-node-1 bash -c sed -i 's/substring/substr/g' /var/lib/clickhouse/metadata/default/file.sql] Executing query BACKUP TABLE tbl, TABLE tbl2 ON CLUSTER 'cluster' TO Disk('backups', '7') on node1 Command:[docker compose --env-file /ClickHouse/tests/integration/test_authentication/_instances-0-gw2/.env --project-name roottestauthentication-gw2 --file /ClickHouse/tests/integration/test_authentication/_instances-0-gw2/instance/docker-compose.yml stop --timeout 20] [gw2] PASSED test_authentication/test.py::test_authentication_pass Connecting to 172.16.7.2(172.16.7.2):2181, use_ssl: False Connection dropped: socket connection error: Connection refused run container_id:roottestattachtablenormalizer-gw9-node-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] Command:[docker exec -u root roottestattachtablenormalizer-gw9-node-1 bash -c ps -C clickhouse] Stdout: PID TTY TIME CMD Stdout: 8 ? 00:00:00 clickhouse run container_id:roottestattachtablenormalizer-gw9-node-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill -9 clickhouse'] Command:[docker exec -u root roottestattachtablenormalizer-gw9-node-1 bash -c pkill -9 clickhouse] Connecting to 172.16.9.4(172.16.9.4):2181, use_ssl: False Connection dropped: socket connection error: Connection refused run container_id:roottestattachtablenormalizer-gw9-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestattachtablenormalizer-gw9-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stderr: Container roottestbackuprestore-gw5-node-1 Stopping Stderr: Container roottestbackuprestore-gw5-node-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_backup_restore/_instances-0-gw5/node/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_backup_restore/_instances-0-gw5/node/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_backup_restore/_instances-0-gw5/.env --project-name roottestbackuprestore-gw5 --file /ClickHouse/tests/integration/test_backup_restore/_instances-0-gw5/node/docker-compose.yml down --volumes] Executing query SYSTEM FLUSH LOGS ON CLUSTER 'cluster' on node1 Stdout:8 run container_id:roottestbackuprestoreoncluster-gw0-node1-1 detach:False nothrow:False cmd: ['bash', '-c', 'grep -a "Writing backup for file .*yPcrXrEIpP.bin .* (disk default)" /var/log/clickhouse-server/clickhouse-server.log | wc -l'] Command:[docker exec roottestbackuprestoreoncluster-gw0-node1-1 bash -c grep -a "Writing backup for file .*yPcrXrEIpP.bin .* (disk default)" /var/log/clickhouse-server/clickhouse-server.log | wc -l] Stdout:1 run container_id:roottestbackuprestoreoncluster-gw0-node2-1 detach:False nothrow:False cmd: ['bash', '-c', 'grep -a "Writing backup for file .*yPcrXrEIpP.bin .* (disk default)" /var/log/clickhouse-server/clickhouse-server.log | wc -l'] Command:[docker exec roottestbackuprestoreoncluster-gw0-node2-1 bash -c grep -a "Writing backup for file .*yPcrXrEIpP.bin .* (disk default)" /var/log/clickhouse-server/clickhouse-server.log | wc -l] Stdout:0 run container_id:roottestbackuprestoreoncluster-gw0-node1-1 detach:False nothrow:False cmd: ['bash', '-c', 'grep -a "Writing backup for file .*yPcrXrEIpP.bin .* skipped" /var/log/clickhouse-server/clickhouse-server.log | wc -l'] Command:[docker exec roottestbackuprestoreoncluster-gw0-node1-1 bash -c grep -a "Writing backup for file .*yPcrXrEIpP.bin .* skipped" /var/log/clickhouse-server/clickhouse-server.log | wc -l] Stdout:3 run container_id:roottestbackuprestoreoncluster-gw0-node2-1 detach:False nothrow:False cmd: ['bash', '-c', 'grep -a "Writing backup for file .*yPcrXrEIpP.bin .* skipped" /var/log/clickhouse-server/clickhouse-server.log | wc -l'] Command:[docker exec roottestbackuprestoreoncluster-gw0-node2-1 bash -c grep -a "Writing backup for file .*yPcrXrEIpP.bin .* skipped" /var/log/clickhouse-server/clickhouse-server.log | wc -l] Stdout:0 Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw0] PASSED test_backup_restore_on_cluster/test.py::test_file_deduplication Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Connecting to 172.16.9.4(172.16.9.4):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Stderr: Container roottestbackuprestore-gw5-node-1 Stopping Stderr: Container roottestbackuprestore-gw5-node-1 Stopped Stderr: Container roottestbackuprestore-gw5-node-1 Removing Stderr: Container roottestbackuprestore-gw5-node-1 Removed Stderr: Network roottestbackuprestore-gw5_default Removing Stderr: Network roottestbackuprestore-gw5_default Removed Cleanup called Docker networks for project roottestbackuprestore-gw5 are NETWORK ID NAME DRIVER SCOPE Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Docker containers for project roottestbackuprestore-gw5 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestbackuprestore-gw5 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestbackuprestore-gw5-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestbackuprestore-gw5 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Connecting to 172.16.7.2(172.16.7.2):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Stdout:1 Volumes pruned: 1 Running tests in /ClickHouse/tests/integration/test_backward_compatibility/test_insert_profile_events.py test_backward_compatibility/test_insert_profile_events.py::test_new_client_compatible Cluster start called. is_up=False Docker networks for project roottestbackwardcompatibilityinsertprofileevents-gw5 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestbackwardcompatibilityinsertprofileevents-gw5 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestbackwardcompatibilityinsertprofileevents-gw5 are DRIVER VOLUME NAME Cleanup called Docker networks for project roottestbackwardcompatibilityinsertprofileevents-gw5 are NETWORK ID NAME DRIVER SCOPE Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Docker containers for project roottestbackwardcompatibilityinsertprofileevents-gw5 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestbackwardcompatibilityinsertprofileevents-gw5 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestbackwardcompatibilityinsertprofileevents-gw5-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestbackwardcompatibilityinsertprofileevents-gw5 Trying to prune unused networks... run container_id:roottestattachtablenormalizer-gw9-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestattachtablenormalizer-gw9-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Executing query CREATE TABLE tbl (`x` UInt8) ENGINE = MergeTree ORDER BY x on node1 test_backup_restore_on_cluster/test.py::test_get_error_from_other_host run container_id:roottestattachtablenormalizer-gw9-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestattachtablenormalizer-gw9-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:1 Volumes pruned: 1 Setup directory for instance: upstream_node Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files [] to /ClickHouse/tests/integration/test_backward_compatibility/_instances-insert_profile_events-0-gw5/upstream_node/configs/config.d Setup database dir /ClickHouse/tests/integration/test_backward_compatibility/_instances-insert_profile_events-0-gw5/upstream_node/database Setup logs dir /ClickHouse/tests/integration/test_backward_compatibility/_instances-insert_profile_events-0-gw5/upstream_node/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Setup directory for instance: old_node Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files [] to /ClickHouse/tests/integration/test_backward_compatibility/_instances-insert_profile_events-0-gw5/old_node/configs/config.d Setup database dir /ClickHouse/tests/integration/test_backward_compatibility/_instances-insert_profile_events-0-gw5/old_node/database Setup logs dir /ClickHouse/tests/integration/test_backward_compatibility/_instances-insert_profile_events-0-gw5/old_node/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw'} stored in /ClickHouse/tests/integration/test_backward_compatibility/_instances-insert_profile_events-0-gw5/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_backward_compatibility/_instances-insert_profile_events-0-gw5/.env --project-name roottestbackwardcompatibilityinsertprofileevents-gw5 --file /ClickHouse/tests/integration/test_backward_compatibility/_instances-insert_profile_events-0-gw5/upstream_node/docker-compose.yml --file /ClickHouse/tests/integration/test_backward_compatibility/_instances-insert_profile_events-0-gw5/old_node/docker-compose.yml pull] Executing query INSERT INTO tbl VALUES (3) on node1 No clickhouse process running. Start new one. http://localhost:None "POST /v1.46/containers/roottestattachtablenormalizer-gw9-node-1/exec HTTP/1.1" 201 74 http://localhost:None "POST /v1.46/exec/99c53e226d92d7d2e302869345e29c19655868efb231d9c86033a501eade1ae6/start HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/exec/99c53e226d92d7d2e302869345e29c19655868efb231d9c86033a501eade1ae6/json HTTP/1.1" 200 586 Executing query SYSTEM FLUSH LOGS on node2 Executing query BACKUP TABLE tbl ON CLUSTER 'cluster' TO Disk('backups', '8') on node1 Executing query SELECT max(value) FROM ( SELECT toStartOfInterval(event_time, toIntervalSecond(1)) AS t, avg(value) AS value FROM system.asynchronous_metric_log WHERE event_time >= now() - 60 AND metric = 'LoadAverage1' GROUP BY t ) SETTINGS max_threads = 1 on node2 Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw0] PASSED test_backup_restore_on_cluster/test.py::test_get_error_from_other_host Executing query SYSTEM FLUSH LOGS on node2 Executing query SELECT max(value) FROM ( SELECT toStartOfInterval(event_time, toIntervalSecond(1)) AS t, avg(value) AS value FROM system.asynchronous_metric_log WHERE event_time >= now() - 60 AND metric = 'OSUptime' GROUP BY t ) SETTINGS max_threads = 1 on node2 Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Executing query SYSTEM FLUSH LOGS on node2 Executing query SELECT max(value) FROM ( SELECT toStartOfInterval(event_time, toIntervalSecond(1)) AS t, avg(value) AS value FROM system.asynchronous_metric_log WHERE event_time >= now() - 60 AND metric = 'OSProcessesRunning' GROUP BY t ) SETTINGS max_threads = 1 on node2 Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Executing query SYSTEM FLUSH LOGS on node2 Executing query SELECT max(value) FROM ( SELECT toStartOfInterval(event_time, toIntervalSecond(1)) AS t, avg(value) AS value FROM system.asynchronous_metric_log WHERE event_time >= now() - 60 AND metric = 'OSInterrupts' GROUP BY t ) SETTINGS max_threads = 1 on node2 Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Executing query SYSTEM FLUSH LOGS on node2 Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Executing query SELECT max(value) FROM ( SELECT toStartOfInterval(event_time, toIntervalSecond(1)) AS t, avg(value) AS value FROM system.asynchronous_metric_log WHERE event_time >= now() - 60 AND metric = 'OSMemoryTotal' GROUP BY t ) SETTINGS max_threads = 1 on node2 [gw6] PASSED test_async_metrics_in_cgroup/test.py::test_system_wide_metrics test_async_metrics_in_cgroup/test.py::test_user_cpu_accounting Executing query SELECT value FROM system.build_options WHERE name = 'CXX_FLAGS' on node1 test_backup_restore_on_cluster/test.py::test_keeper_value_max_size Executing query CREATE TABLE tbl ON CLUSTER 'cluster' (x UInt32) ENGINE=ReplicatedMergeTree('/clickhouse/tables/tbl/', '{replica}')ORDER BY x on node1 Executing query SELECT sum(*) FROM system.numbers_mt FORMAT Null SETTINGS max_execution_time=10 on node2 Connecting to 172.16.7.2(172.16.7.2):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost Connecting to 172.16.9.4(172.16.9.4):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost Executing query INSERT INTO tbl VALUES (111) on node1 run container_id:roottestattachtablenormalizer-gw9-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestattachtablenormalizer-gw9-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED get_instance_ip instance_name=zoo2 http://localhost:None "GET /v1.46/containers/roottestalteronmixedtypecluster-gw7-zoo2-1/json HTTP/1.1" 200 None get_kazoo_client: zoo2, ip:172.16.7.3, port:2181, use_ssl:False Connecting to 172.16.7.3(172.16.7.3):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost Executing query INSERT INTO tbl VALUES (222) on node2 Stdout:741 Clickhouse process running. run container_id:roottestattachtablenormalizer-gw9-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestattachtablenormalizer-gw9-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Connecting to 172.16.2.3(172.16.2.3):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED get_instance_ip instance_name=zoo3 Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster' tbl on node1 http://localhost:None "GET /v1.46/containers/roottestalteronmixedtypecluster-gw7-zoo3-1/json HTTP/1.1" 200 None get_kazoo_client: zoo3, ip:172.16.7.4, port:2181, use_ssl:False Connecting to 172.16.7.4(172.16.7.4):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost Stdout:741 Executing query select 20 on node Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED All instances of ZooKeeper started: ('zoo1', 'zoo2', 'zoo3') ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_alter_on_mixed_type_cluster/_instances-0-gw7/.env --project-name roottestalteronmixedtypecluster-gw7 --file /ClickHouse/tests/integration/test_alter_on_mixed_type_cluster/_instances-0-gw7/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_alter_on_mixed_type_cluster/_instances-0-gw7/node2/docker-compose.yml --file /ClickHouse/tests/integration/test_alter_on_mixed_type_cluster/_instances-0-gw7/node3/docker-compose.yml --file /ClickHouse/tests/integration/test_alter_on_mixed_type_cluster/_instances-0-gw7/node4/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_alter_on_mixed_type_cluster/_instances-0-gw7/.env --project-name roottestalteronmixedtypecluster-gw7 --file /ClickHouse/tests/integration/test_alter_on_mixed_type_cluster/_instances-0-gw7/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_alter_on_mixed_type_cluster/_instances-0-gw7/node2/docker-compose.yml --file /ClickHouse/tests/integration/test_alter_on_mixed_type_cluster/_instances-0-gw7/node3/docker-compose.yml --file /ClickHouse/tests/integration/test_alter_on_mixed_type_cluster/_instances-0-gw7/node4/docker-compose.yml up -d --no-recreate] Connecting to 172.16.2.3(172.16.2.3):2181, use_ssl: False Connection dropped: socket connection error: Connection refused [gw9] PASSED test_attach_table_normalizer/test.py::test_attach_substr_restart Command:[docker compose --env-file /ClickHouse/tests/integration/test_attach_table_normalizer/_instances-0-gw9/.env --project-name roottestattachtablenormalizer-gw9 --file /ClickHouse/tests/integration/test_attach_table_normalizer/_instances-0-gw9/node/docker-compose.yml stop --timeout 20] Stderr: Container roottestauthentication-gw2-instance-1 Stopping Stderr: Container roottestauthentication-gw2-instance-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_authentication/_instances-0-gw2/instance/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_authentication/_instances-0-gw2/instance/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_authentication/_instances-0-gw2/.env --project-name roottestauthentication-gw2 --file /ClickHouse/tests/integration/test_authentication/_instances-0-gw2/instance/docker-compose.yml down --volumes] Executing query SYSTEM STOP REPLICATED SENDS ON CLUSTER 'cluster' tbl on node1 Connection dropped: outstanding heartbeat ping not received Transition to CONNECTING Zookeeper connection lost Executing query INSERT INTO tbl VALUES (333) on node1 Connecting to 172.16.2.3(172.16.2.3):2181, use_ssl: False Connecting to 172.16.2.3(172.16.2.3):2181, use_ssl: False Executing query INSERT INTO tbl VALUES (444) on node2 Executing query BACKUP TABLE tbl ON CLUSTER 'cluster' TO Disk('backups', '9') on node1 Stderr: Container roottestattachtablenormalizer-gw9-node-1 Stopping Stderr: Container roottestattachtablenormalizer-gw9-node-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_attach_table_normalizer/_instances-0-gw9/node/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_attach_table_normalizer/_instances-0-gw9/node/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_attach_table_normalizer/_instances-0-gw9/.env --project-name roottestattachtablenormalizer-gw9 --file /ClickHouse/tests/integration/test_attach_table_normalizer/_instances-0-gw9/node/docker-compose.yml down --volumes] Stderr: Container roottestauthentication-gw2-instance-1 Stopping Stderr: Container roottestauthentication-gw2-instance-1 Stopped Stderr: Container roottestauthentication-gw2-instance-1 Removing Stderr: Container roottestauthentication-gw2-instance-1 Removed Stderr: Network roottestauthentication-gw2_default Removing Stderr: Network roottestauthentication-gw2_default Removed Cleanup called Docker networks for project roottestauthentication-gw2 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestauthentication-gw2 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestauthentication-gw2 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestauthentication-gw2-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestauthentication-gw2 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:1 Volumes pruned: 1 Running tests in /ClickHouse/tests/integration/test_alternative_keeper_config/test.py test_alternative_keeper_config/test.py::test_create_insert Cluster start called. is_up=False Docker networks for project roottestalternativekeeperconfig-gw2 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestalternativekeeperconfig-gw2 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestalternativekeeperconfig-gw2 are DRIVER VOLUME NAME Cleanup called Docker networks for project roottestalternativekeeperconfig-gw2 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestalternativekeeperconfig-gw2 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestalternativekeeperconfig-gw2 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestalternativekeeperconfig-gw2-.*-1$' --format '{{.ID}}:{{.Names}}'] Stderr: Container roottestalteronmixedtypecluster-gw7-zoo3-1 Running Stderr: Container roottestalteronmixedtypecluster-gw7-zoo1-1 Running Stderr: Container roottestalteronmixedtypecluster-gw7-zoo2-1 Running Stderr: Container roottestalteronmixedtypecluster-gw7-node1-1 Creating Stderr: Container roottestalteronmixedtypecluster-gw7-node2-1 Creating Stderr: Container roottestalteronmixedtypecluster-gw7-node3-1 Creating Stderr: Container roottestalteronmixedtypecluster-gw7-node4-1 Creating Stderr: Container roottestalteronmixedtypecluster-gw7-node4-1 Created Stderr: Container roottestalteronmixedtypecluster-gw7-node3-1 Created Stderr: Container roottestalteronmixedtypecluster-gw7-node2-1 Created Stderr: Container roottestalteronmixedtypecluster-gw7-node1-1 Created Stderr: Container roottestalteronmixedtypecluster-gw7-node4-1 Starting Stderr: Container roottestalteronmixedtypecluster-gw7-node1-1 Starting Stderr: Container roottestalteronmixedtypecluster-gw7-node2-1 Starting Stderr: Container roottestalteronmixedtypecluster-gw7-node3-1 Starting Stderr: Container roottestalteronmixedtypecluster-gw7-node4-1 Started Stderr: Container roottestalteronmixedtypecluster-gw7-node1-1 Started Stderr: Container roottestalteronmixedtypecluster-gw7-node3-1 Started Stderr: Container roottestalteronmixedtypecluster-gw7-node2-1 Started ClickHouse instance created get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottestalteronmixedtypecluster-gw7-node1-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottestalteronmixedtypecluster-gw7-node1-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node1, ip: 172.16.7.5... http://localhost:None "GET /v1.46/containers/roottestalteronmixedtypecluster-gw7-node1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/2396383ab7eb7549494dbc9d7470fb81ff5828f2010a4f0fd7fadc7ff8e7652b/json HTTP/1.1" 200 None Unstopped containers: {} No running containers for project: roottestalternativekeeperconfig-gw2 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:1 Volumes pruned: 1 Setup directory for instance: node1 Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_alternative_keeper_config/configs/remote_servers.xml', '/ClickHouse/tests/integration/test_alternative_keeper_config/configs/keeper_config.xml', '/ClickHouse/tests/integration/test_alternative_keeper_config/configs/enable_keeper1.xml'] to /ClickHouse/tests/integration/test_alternative_keeper_config/_instances-0-gw2/node1/configs/config.d Setup database dir /ClickHouse/tests/integration/test_alternative_keeper_config/_instances-0-gw2/node1/database Setup logs dir /ClickHouse/tests/integration/test_alternative_keeper_config/_instances-0-gw2/node1/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Setup directory for instance: node2 Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_alternative_keeper_config/configs/remote_servers.xml', '/ClickHouse/tests/integration/test_alternative_keeper_config/configs/zookeeper_config.xml', '/ClickHouse/tests/integration/test_alternative_keeper_config/configs/enable_keeper2.xml'] to /ClickHouse/tests/integration/test_alternative_keeper_config/_instances-0-gw2/node2/configs/config.d Setup database dir /ClickHouse/tests/integration/test_alternative_keeper_config/_instances-0-gw2/node2/database Setup logs dir /ClickHouse/tests/integration/test_alternative_keeper_config/_instances-0-gw2/node2/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Setup directory for instance: node3 Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_alternative_keeper_config/configs/remote_servers.xml', '/ClickHouse/tests/integration/test_alternative_keeper_config/configs/enable_keeper3.xml'] to /ClickHouse/tests/integration/test_alternative_keeper_config/_instances-0-gw2/node3/configs/config.d Setup database dir /ClickHouse/tests/integration/test_alternative_keeper_config/_instances-0-gw2/node3/database Setup logs dir /ClickHouse/tests/integration/test_alternative_keeper_config/_instances-0-gw2/node3/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw'} stored in /ClickHouse/tests/integration/test_alternative_keeper_config/_instances-0-gw2/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found http://localhost:None "GET /version HTTP/1.1" 200 826 http://localhost:None "GET /v1.46/containers/2396383ab7eb7549494dbc9d7470fb81ff5828f2010a4f0fd7fadc7ff8e7652b/json HTTP/1.1" 200 None Command:[docker compose --env-file /ClickHouse/tests/integration/test_alternative_keeper_config/_instances-0-gw2/.env --project-name roottestalternativekeeperconfig-gw2 --file /ClickHouse/tests/integration/test_alternative_keeper_config/_instances-0-gw2/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_alternative_keeper_config/_instances-0-gw2/node2/docker-compose.yml --file /ClickHouse/tests/integration/test_alternative_keeper_config/_instances-0-gw2/node3/docker-compose.yml pull] http://localhost:None "GET /v1.46/containers/2396383ab7eb7549494dbc9d7470fb81ff5828f2010a4f0fd7fadc7ff8e7652b/json HTTP/1.1" 200 None Stderr: Container roottestattachtablenormalizer-gw9-node-1 Stopping Stderr: Container roottestattachtablenormalizer-gw9-node-1 Stopped Stderr: Container roottestattachtablenormalizer-gw9-node-1 Removing Stderr: Container roottestattachtablenormalizer-gw9-node-1 Removed Stderr: Network roottestattachtablenormalizer-gw9_default Removing Stderr: Network roottestattachtablenormalizer-gw9_default Removed Cleanup called Docker networks for project roottestattachtablenormalizer-gw9 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestattachtablenormalizer-gw9 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Executing query DROP TABLE tbl ON CLUSTER 'cluster' SYNC on node1 Docker volumes for project roottestattachtablenormalizer-gw9 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestattachtablenormalizer-gw9-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestattachtablenormalizer-gw9 Trying to prune unused networks... http://localhost:None "GET /v1.46/containers/2396383ab7eb7549494dbc9d7470fb81ff5828f2010a4f0fd7fadc7ff8e7652b/json HTTP/1.1" 200 None Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:1 Volumes pruned: 1 test_backward_compatibility/test_functions.py::test_aggregate_states Running tests in /ClickHouse/tests/integration/test_backward_compatibility/test_functions.py Cluster start called. is_up=False Docker networks for project roottestbackwardcompatibilityfunctions-gw9 are NETWORK ID NAME DRIVER SCOPE http://localhost:None "GET /v1.46/containers/2396383ab7eb7549494dbc9d7470fb81ff5828f2010a4f0fd7fadc7ff8e7652b/json HTTP/1.1" 200 None Executing query RESTORE TABLE tbl ON CLUSTER 'cluster' FROM Disk('backups', '9') on node1 Docker containers for project roottestbackwardcompatibilityfunctions-gw9 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Connecting to 172.16.9.4(172.16.9.4):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost Docker volumes for project roottestbackwardcompatibilityfunctions-gw9 are DRIVER VOLUME NAME Cleanup called Docker networks for project roottestbackwardcompatibilityfunctions-gw9 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestbackwardcompatibilityfunctions-gw9 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestbackwardcompatibilityfunctions-gw9 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestbackwardcompatibilityfunctions-gw9-.*-1$' --format '{{.ID}}:{{.Names}}'] Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED get_instance_ip instance_name=zoo2 http://localhost:None "GET /v1.46/containers/roottestattachpartitionwithlargedestination-gw8-zoo2-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/2396383ab7eb7549494dbc9d7470fb81ff5828f2010a4f0fd7fadc7ff8e7652b/json HTTP/1.1" 200 None get_kazoo_client: zoo2, ip:172.16.9.3, port:2181, use_ssl:False Connecting to 172.16.9.3(172.16.9.3):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Unstopped containers: {} No running containers for project: roottestbackwardcompatibilityfunctions-gw9 Trying to prune unused networks... Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:1 Volumes pruned: 1 Setup directory for instance: upstream Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files [] to /ClickHouse/tests/integration/test_backward_compatibility/_instances-functions-0-gw9/upstream/configs/config.d Setup database dir /ClickHouse/tests/integration/test_backward_compatibility/_instances-functions-0-gw9/upstream/database Setup logs dir /ClickHouse/tests/integration/test_backward_compatibility/_instances-functions-0-gw9/upstream/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Setup directory for instance: backward Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files [] to /ClickHouse/tests/integration/test_backward_compatibility/_instances-functions-0-gw9/backward/configs/config.d Setup database dir /ClickHouse/tests/integration/test_backward_compatibility/_instances-functions-0-gw9/backward/database Setup logs dir /ClickHouse/tests/integration/test_backward_compatibility/_instances-functions-0-gw9/backward/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw'} stored in /ClickHouse/tests/integration/test_backward_compatibility/_instances-functions-0-gw9/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED get_instance_ip instance_name=zoo3 http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_backward_compatibility/_instances-functions-0-gw9/.env --project-name roottestbackwardcompatibilityfunctions-gw9 --file /ClickHouse/tests/integration/test_backward_compatibility/_instances-functions-0-gw9/upstream/docker-compose.yml --file /ClickHouse/tests/integration/test_backward_compatibility/_instances-functions-0-gw9/backward/docker-compose.yml pull] http://localhost:None "GET /v1.46/containers/roottestattachpartitionwithlargedestination-gw8-zoo3-1/json HTTP/1.1" 200 None get_kazoo_client: zoo3, ip:172.16.9.2, port:2181, use_ssl:False http://localhost:None "GET /v1.46/containers/2396383ab7eb7549494dbc9d7470fb81ff5828f2010a4f0fd7fadc7ff8e7652b/json HTTP/1.1" 200 None ClickHouse node1 started get_instance_ip instance_name=node2 Connecting to 172.16.9.2(172.16.9.2):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) http://localhost:None "GET /v1.46/containers/roottestalteronmixedtypecluster-gw7-node2-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node2 http://localhost:None "GET /v1.46/containers/roottestalteronmixedtypecluster-gw7-node2-1/json HTTP/1.1" 200 None Zookeeper connection established, state: CONNECTED Waiting for ClickHouse start in node2, ip: 172.16.7.8... http://localhost:None "GET /v1.46/containers/roottestalteronmixedtypecluster-gw7-node2-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/07881ec1be35c197e4a2e168cc3c330ab7906e0c17f21a13eeba7f8cb1578ce0/json HTTP/1.1" 200 None Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster' tbl on node1 http://localhost:None "GET /v1.46/containers/07881ec1be35c197e4a2e168cc3c330ab7906e0c17f21a13eeba7f8cb1578ce0/json HTTP/1.1" 200 None ClickHouse node2 started get_instance_ip instance_name=node3 http://localhost:None "GET /v1.46/containers/roottestalteronmixedtypecluster-gw7-node3-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node3 Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED All instances of ZooKeeper started: ('zoo1', 'zoo2', 'zoo3') ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_attach_partition_with_large_destination/_instances-0-gw8/.env --project-name roottestattachpartitionwithlargedestination-gw8 --file /ClickHouse/tests/integration/test_attach_partition_with_large_destination/_instances-0-gw8/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_attach_partition_with_large_destination/_instances-0-gw8/.env --project-name roottestattachpartitionwithlargedestination-gw8 --file /ClickHouse/tests/integration/test_attach_partition_with_large_destination/_instances-0-gw8/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml up -d --no-recreate] http://localhost:None "GET /v1.46/containers/roottestalteronmixedtypecluster-gw7-node3-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node3, ip: 172.16.7.7... http://localhost:None "GET /v1.46/containers/roottestalteronmixedtypecluster-gw7-node3-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/5bb948aede8b002652c471b45f8ba00a8c49ab5f0a36df10da2ba1652b063ef6/json HTTP/1.1" 200 None ClickHouse node3 started get_instance_ip instance_name=node4 http://localhost:None "GET /v1.46/containers/roottestalteronmixedtypecluster-gw7-node4-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node4 http://localhost:None "GET /v1.46/containers/roottestalteronmixedtypecluster-gw7-node4-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node4, ip: 172.16.7.6... http://localhost:None "GET /v1.46/containers/roottestalteronmixedtypecluster-gw7-node4-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/34f1c84c7bf909abc5de4e629299c53cd113077fb1740aa4e9538b9e05d7aef2/json HTTP/1.1" 200 None ClickHouse node4 started Executing query CREATE TABLE IF NOT EXISTS test_table_replicated(date Date, id UInt32, value Int32) ENGINE = ReplicatedMergeTree('/clickhouse/tables/0/sometable', 'node1') ORDER BY id; on node1 Executing query SELECT * FROM tbl ORDER BY x on node1 Executing query CREATE TABLE IF NOT EXISTS test_table(date Date, id UInt32, value Int32) ENGINE=MergeTree ORDER BY id on node1 Executing query CREATE TABLE IF NOT EXISTS test_table_replicated(date Date, id UInt32, value Int32) ENGINE = ReplicatedMergeTree('/clickhouse/tables/0/sometable', 'node2') ORDER BY id; on node2 Executing query SELECT * FROM tbl ORDER BY x on node2 Executing query CREATE TABLE IF NOT EXISTS test_table(date Date, id UInt32, value Int32) ENGINE=MergeTree ORDER BY id on node2 Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw0] PASSED test_backup_restore_on_cluster/test.py::test_keeper_value_max_size Executing query CREATE TABLE IF NOT EXISTS test_table_replicated(date Date, id UInt32, value Int32) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/someotable', 'node3') ORDER BY id; on node3 Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Stderr: Container roottestattachpartitionwithlargedestination-gw8-zoo2-1 Running Stderr: Container roottestattachpartitionwithlargedestination-gw8-zoo3-1 Running Stderr: Container roottestattachpartitionwithlargedestination-gw8-zoo1-1 Running Stderr: Container roottestattachpartitionwithlargedestination-gw8-node-1 Creating Stderr: Container roottestattachpartitionwithlargedestination-gw8-node-1 Created Stderr: Container roottestattachpartitionwithlargedestination-gw8-node-1 Starting Stderr: Container roottestattachpartitionwithlargedestination-gw8-node-1 Started ClickHouse instance created get_instance_ip instance_name=node http://localhost:None "GET /v1.46/containers/roottestattachpartitionwithlargedestination-gw8-node-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node http://localhost:None "GET /v1.46/containers/roottestattachpartitionwithlargedestination-gw8-node-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node, ip: 172.16.9.5... http://localhost:None "GET /v1.46/containers/roottestattachpartitionwithlargedestination-gw8-node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/0a63880beaa4a409d6d0b6aa1c44a74ab812834f4a5d8bfde7028503e4d5b7d1/json HTTP/1.1" 200 None Executing query CREATE TABLE IF NOT EXISTS test_table(date Date, id UInt32, value Int32) ENGINE=MergeTree ORDER BY id on node3 Executing query CREATE TABLE IF NOT EXISTS test_table_replicated(date Date, id UInt32, value Int32) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/someotable', 'node4') ORDER BY id; on node4 http://localhost:None "GET /v1.46/containers/0a63880beaa4a409d6d0b6aa1c44a74ab812834f4a5d8bfde7028503e4d5b7d1/json HTTP/1.1" 200 None Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Executing query CREATE TABLE IF NOT EXISTS test_table(date Date, id UInt32, value Int32) ENGINE=MergeTree ORDER BY id on node4 http://localhost:None "GET /v1.46/containers/0a63880beaa4a409d6d0b6aa1c44a74ab812834f4a5d8bfde7028503e4d5b7d1/json HTTP/1.1" 200 None Executing query DROP TABLE IF EXISTS u; on node1 [gw1] PASSED test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[two_shards_three_nodes] test_cluster_all_replicas/test.py::test_global_in Executing query CREATE TABLE u(uid Int16) ENGINE=Memory as select 0 on node1 Executing query INSERT INTO test_table VALUES(toDate('2019-10-01'), 1, 1) on node1 Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 http://localhost:None "GET /v1.46/containers/0a63880beaa4a409d6d0b6aa1c44a74ab812834f4a5d8bfde7028503e4d5b7d1/json HTTP/1.1" 200 None Executing query INSERT INTO test_table VALUES(toDate('2019-10-01'), 1, 1) on node2 Executing query SELECT hostName(), * FROM clusterAllReplicas("one_shard_two_nodes", system.one) where dummy GLOBAL IN u ORDER BY ALL on node1 Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Executing query INSERT INTO test_table VALUES(toDate('2019-10-01'), 1, 1) on node3 http://localhost:None "GET /v1.46/containers/0a63880beaa4a409d6d0b6aa1c44a74ab812834f4a5d8bfde7028503e4d5b7d1/json HTTP/1.1" 200 None Executing query SELECT hostName() FROM clusterAllReplicas('one_shard_three_nodes', system.one) ORDER BY ALL settings skip_unavailable_shards=1 on node1 [gw1] PASSED test_cluster_all_replicas/test.py::test_global_in test_cluster_all_replicas/test.py::test_skip_unavailable_replica[one_shard_three_nodes] Executing query INSERT INTO test_table VALUES(toDate('2019-10-01'), 1, 1) on node4 Executing query CREATE TABLE tbl ON CLUSTER 'cluster' (x UInt8, y String) ENGINE=ReplicatedMergeTree('/clickhouse/tables/tbl/', '{replica}')ORDER BY tuple() on node1 http://localhost:None "GET /v1.46/containers/0a63880beaa4a409d6d0b6aa1c44a74ab812834f4a5d8bfde7028503e4d5b7d1/json HTTP/1.1" 200 None ClickHouse node started test_backup_restore_on_cluster/test.py::test_mutation Executing query CREATE DATABASE db ENGINE=Atomic on node Executing query SELECT COUNT() FROM test_table on node1 Executing query CREATE TABLE db.destination (n UInt64) ENGINE=ReplicatedMergeTree('/test/destination', 'r1') ORDER BY n PARTITION BY n % 2 on node Executing query CREATE TABLE db.source_1 (n UInt64) ENGINE=ReplicatedMergeTree('/test/source_1', 'r1') ORDER BY n PARTITION BY n % 2 on node Executing query SELECT COUNT() FROM test_table on node2 Executing query INSERT INTO tbl SELECT number, toString(number) FROM numbers(5) on node1 Executing query INSERT INTO db.source_1 VALUES (1), (2), (3), (4) on node Executing query INSERT INTO tbl SELECT number, toString(number) FROM numbers(5, 5) on node2 Executing query SELECT COUNT() FROM test_table on node3 Executing query INSERT INTO tbl SELECT number, toString(number) FROM numbers(10, 5) on node1 Executing query CREATE TABLE db.source_2 (n UInt64) ENGINE=ReplicatedMergeTree('/test/source_2', 'r1') ORDER BY n PARTITION BY n % 2 on node Executing query ALTER TABLE tbl UPDATE x=x+1 WHERE 1 on node1 Executing query INSERT INTO db.source_2 VALUES (5), (6), (7), (8) on node Executing query SELECT COUNT() FROM test_table on node4 Executing query ALTER TABLE tbl UPDATE x=x+1+sleep(3) WHERE 1 on node1 Executing query ALTER TABLE db.destination ATTACH PARTITION 0 FROM db.source_1 on node Executing query ALTER TABLE tbl UPDATE x=x+1+sleep(3) WHERE 1 on node1 Executing query SELECT n FROM db.destination ORDER BY n on node Executing query ALTER TABLE test_table ON CLUSTER 'test_cluster_mixed' MODIFY COLUMN date DateTime on node1 Executing query BACKUP TABLE tbl ON CLUSTER 'cluster' TO Disk('backups', '10') on node1 Executing query ALTER TABLE db.destination REPLACE PARTITION 0 FROM db.source_2 on node Executing query SELECT n FROM db.destination ORDER BY n on node Executing query SELECT date FROM test_table on node1 Executing query ALTER TABLE db.destination ATTACH PARTITION 0 FROM db.source_2 on node Executing query SELECT date FROM test_table on node2 Executing query SELECT n FROM db.destination ORDER BY n on node Executing query SELECT date FROM test_table on node3 run container_id:roottestattachpartitionwithlargedestination-gw8-node-1 detach:False nothrow:False cmd: ['bash', '-c', 'touch /var/lib/clickhouse/flags/force_drop_table && chmod a=rw /var/lib/clickhouse/flags/force_drop_table'] Command:[docker exec -u root roottestattachpartitionwithlargedestination-gw8-node-1 bash -c touch /var/lib/clickhouse/flags/force_drop_table && chmod a=rw /var/lib/clickhouse/flags/force_drop_table] Executing query DROP TABLE tbl ON CLUSTER 'cluster' SYNC on node1 Executing query SELECT date FROM test_table on node4 Executing query DROP TABLE db.source_1 SYNC on node run container_id:roottestattachpartitionwithlargedestination-gw8-node-1 detach:False nothrow:False cmd: ['bash', '-c', 'touch /var/lib/clickhouse/flags/force_drop_table && chmod a=rw /var/lib/clickhouse/flags/force_drop_table'] Command:[docker exec -u root roottestattachpartitionwithlargedestination-gw8-node-1 bash -c touch /var/lib/clickhouse/flags/force_drop_table && chmod a=rw /var/lib/clickhouse/flags/force_drop_table] Executing query ALTER TABLE test_table ON CLUSTER 'test_cluster_mixed' MODIFY COLUMN value String on node3 Executing query DROP TABLE db.source_2 SYNC on node Executing query INSERT INTO test_table VALUES(toDateTime('2019-10-02 00:00:00'), 2, 'Hello') on node1 run container_id:roottestattachpartitionwithlargedestination-gw8-node-1 detach:False nothrow:False cmd: ['bash', '-c', 'touch /var/lib/clickhouse/flags/force_drop_table && chmod a=rw /var/lib/clickhouse/flags/force_drop_table'] Command:[docker exec -u root roottestattachpartitionwithlargedestination-gw8-node-1 bash -c touch /var/lib/clickhouse/flags/force_drop_table && chmod a=rw /var/lib/clickhouse/flags/force_drop_table] Executing query INSERT INTO test_table VALUES(toDateTime('2019-10-02 00:00:00'), 2, 'Hello') on node2 Executing query DROP TABLE db.destination SYNC on node Executing query INSERT INTO test_table VALUES(toDateTime('2019-10-02 00:00:00'), 2, 'Hello') on node3 Executing query DROP DATABASE db on node Executing query INSERT INTO test_table VALUES(toDateTime('2019-10-02 00:00:00'), 2, 'Hello') on node4 [gw8] PASSED test_attach_partition_with_large_destination/test.py::test_attach_partition_with_large_destination[Atomic] Executing query CREATE DATABASE db ENGINE=Ordinary on node test_attach_partition_with_large_destination/test.py::test_attach_partition_with_large_destination[Ordinary] Executing query SELECT COUNT() FROM test_table on node1 Executing query CREATE TABLE db.destination (n UInt64) ENGINE=ReplicatedMergeTree('/test/destination', 'r1') ORDER BY n PARTITION BY n % 2 on node Executing query SELECT COUNT() FROM test_table on node2 Executing query CREATE TABLE db.source_1 (n UInt64) ENGINE=ReplicatedMergeTree('/test/source_1', 'r1') ORDER BY n PARTITION BY n % 2 on node Executing query SELECT COUNT() FROM test_table on node3 Executing query INSERT INTO db.source_1 VALUES (1), (2), (3), (4) on node Connection dropped: socket connection error: No route to host Connection dropped: socket connection error: No route to host Executing query SELECT COUNT() FROM test_table on node4 Executing query CREATE TABLE db.source_2 (n UInt64) ENGINE=ReplicatedMergeTree('/test/source_2', 'r1') ORDER BY n PARTITION BY n % 2 on node Executing query TRUNCATE TABLE test_table on node1 Executing query INSERT INTO db.source_2 VALUES (5), (6), (7), (8) on node Executing query TRUNCATE TABLE test_table on node2 Executing query ALTER TABLE db.destination ATTACH PARTITION 0 FROM db.source_1 on node Executing query TRUNCATE TABLE test_table on node3 Executing query SELECT n FROM db.destination ORDER BY n on node Executing query TRUNCATE TABLE test_table on node4 Executing query ALTER TABLE db.destination REPLACE PARTITION 0 FROM db.source_2 on node Connecting to 172.16.2.3(172.16.2.3):2181, use_ssl: False Executing query INSERT INTO test_table_replicated VALUES(toDate('2019-10-01'), 1, 1) on node1 [gw7] PASSED test_alter_on_mixed_type_cluster/test.py::test_alter_on_cluter_non_replicated test_alter_on_mixed_type_cluster/test.py::test_alter_replicated_on_cluster Executing query SELECT n FROM db.destination ORDER BY n on node Executing query INSERT INTO test_table_replicated VALUES(toDate('2019-10-01'), 1, 1) on node3 Executing query ALTER TABLE db.destination ATTACH PARTITION 0 FROM db.source_2 on node Executing query SYSTEM SYNC REPLICA test_table_replicated on node2 Executing query SELECT n FROM db.destination ORDER BY n on node Executing query SYSTEM SYNC REPLICA test_table_replicated on node4 run container_id:roottestattachpartitionwithlargedestination-gw8-node-1 detach:False nothrow:False cmd: ['bash', '-c', 'touch /var/lib/clickhouse/flags/force_drop_table && chmod a=rw /var/lib/clickhouse/flags/force_drop_table'] Command:[docker exec -u root roottestattachpartitionwithlargedestination-gw8-node-1 bash -c touch /var/lib/clickhouse/flags/force_drop_table && chmod a=rw /var/lib/clickhouse/flags/force_drop_table] Executing query ALTER TABLE test_table_replicated ON CLUSTER 'test_cluster_mixed' MODIFY COLUMN date DateTime on node1 Executing query DROP TABLE db.source_1 SYNC on node Stderr: Container roottestconfigsubstitutions-gw3-node3-1 Stopping Stderr: Container roottestconfigsubstitutions-gw3-node2-1 Stopping Stderr: Container roottestconfigsubstitutions-gw3-node5-1 Stopping Stderr: Container roottestconfigsubstitutions-gw3-node1-1 Stopping Stderr: Container roottestconfigsubstitutions-gw3-node4-1 Stopping Stderr: Container roottestconfigsubstitutions-gw3-node7-1 Stopping Stderr: Container roottestconfigsubstitutions-gw3-node6-1 Stopping Stderr: Container roottestconfigsubstitutions-gw3-node8-1 Stopping Stderr: Container roottestconfigsubstitutions-gw3-node3-1 Stopped Stderr: Container roottestconfigsubstitutions-gw3-zoo3-1 Stopping Stderr: Container roottestconfigsubstitutions-gw3-zoo1-1 Stopping Stderr: Container roottestconfigsubstitutions-gw3-zoo2-1 Stopping Stderr: Container roottestconfigsubstitutions-gw3-zoo2-1 Stopped Stderr: Container roottestconfigsubstitutions-gw3-zoo3-1 Stopped Stderr: Container roottestconfigsubstitutions-gw3-node8-1 Stopped Stderr: Container roottestconfigsubstitutions-gw3-node7-1 Stopped Stderr: Container roottestconfigsubstitutions-gw3-node1-1 Stopped Stderr: Container roottestconfigsubstitutions-gw3-node2-1 Stopped Stderr: Container roottestconfigsubstitutions-gw3-node5-1 Stopped Stderr: Container roottestconfigsubstitutions-gw3-node4-1 Stopped Stderr: Container roottestconfigsubstitutions-gw3-node6-1 Stopped Stderr: Container roottestconfigsubstitutions-gw3-zoo1-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node1/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node1/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node2/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node2/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] run container_id:roottestattachpartitionwithlargedestination-gw8-node-1 detach:False nothrow:False cmd: ['bash', '-c', 'touch /var/lib/clickhouse/flags/force_drop_table && chmod a=rw /var/lib/clickhouse/flags/force_drop_table'] Command:[docker exec -u root roottestattachpartitionwithlargedestination-gw8-node-1 bash -c touch /var/lib/clickhouse/flags/force_drop_table && chmod a=rw /var/lib/clickhouse/flags/force_drop_table] Command:[bash -c [ -f /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node3/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node3/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node4/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node4/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Executing query DROP TABLE db.source_2 SYNC on node Command:[bash -c [ -f /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node5/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node5/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node6/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node6/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node7/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node7/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node8/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node8/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] run container_id:roottestattachpartitionwithlargedestination-gw8-node-1 detach:False nothrow:False cmd: ['bash', '-c', 'touch /var/lib/clickhouse/flags/force_drop_table && chmod a=rw /var/lib/clickhouse/flags/force_drop_table'] Command:[docker exec -u root roottestattachpartitionwithlargedestination-gw8-node-1 bash -c touch /var/lib/clickhouse/flags/force_drop_table && chmod a=rw /var/lib/clickhouse/flags/force_drop_table] Executing query SELECT date FROM test_table_replicated on node1 Connecting to 172.16.2.3(172.16.2.3):2181, use_ssl: False Command:[docker compose --env-file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/.env --project-name roottestconfigsubstitutions-gw3 --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node2/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node3/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node4/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node5/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node6/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node7/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw3/node8/docker-compose.yml down --volumes] Executing query DROP TABLE db.destination SYNC on node Executing query SELECT date FROM test_table_replicated on node2 Executing query SELECT date FROM test_table_replicated on node3 Executing query DROP DATABASE db on node Executing query SELECT date FROM test_table_replicated on node4 Command:[docker compose --env-file /ClickHouse/tests/integration/test_attach_partition_with_large_destination/_instances-0-gw8/.env --project-name roottestattachpartitionwithlargedestination-gw8 --file /ClickHouse/tests/integration/test_attach_partition_with_large_destination/_instances-0-gw8/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml stop --timeout 20] [gw8] PASSED test_attach_partition_with_large_destination/test.py::test_attach_partition_with_large_destination[Ordinary] Executing query ALTER TABLE test_table_replicated ON CLUSTER 'test_cluster_mixed' MODIFY COLUMN value String on node3 Executing query INSERT INTO test_table_replicated VALUES(toDateTime('2019-10-02 00:00:00'), 2, 'Hello') on node2 Executing query INSERT INTO test_table_replicated VALUES(toDateTime('2019-10-02 00:00:00'), 2, 'Hello') on node4 Executing query SYSTEM SYNC REPLICA test_table_replicated on node1 Executing query SYSTEM SYNC REPLICA test_table_replicated on node3 Stderr: Container roottestconfigsubstitutions-gw3-node4-1 Stopping Stderr: Container roottestconfigsubstitutions-gw3-node1-1 Stopping Stderr: Container roottestconfigsubstitutions-gw3-node3-1 Stopping Stderr: Container roottestconfigsubstitutions-gw3-node8-1 Stopping Stderr: Container roottestconfigsubstitutions-gw3-node5-1 Stopping Stderr: Container roottestconfigsubstitutions-gw3-node7-1 Stopping Stderr: Container roottestconfigsubstitutions-gw3-node6-1 Stopping Stderr: Container roottestconfigsubstitutions-gw3-node2-1 Stopping Stderr: Container roottestconfigsubstitutions-gw3-node1-1 Stopped Stderr: Container roottestconfigsubstitutions-gw3-node1-1 Removing Stderr: Container roottestconfigsubstitutions-gw3-node4-1 Stopped Stderr: Container roottestconfigsubstitutions-gw3-node4-1 Removing Stderr: Container roottestconfigsubstitutions-gw3-node2-1 Stopped Stderr: Container roottestconfigsubstitutions-gw3-node2-1 Removing Stderr: Container roottestconfigsubstitutions-gw3-node3-1 Stopped Stderr: Container roottestconfigsubstitutions-gw3-node3-1 Removing Stderr: Container roottestconfigsubstitutions-gw3-node5-1 Stopped Stderr: Container roottestconfigsubstitutions-gw3-node5-1 Removing Stderr: Container roottestconfigsubstitutions-gw3-node7-1 Stopped Stderr: Container roottestconfigsubstitutions-gw3-node7-1 Removing Stderr: Container roottestconfigsubstitutions-gw3-node8-1 Stopped Stderr: Container roottestconfigsubstitutions-gw3-node8-1 Removing Stderr: Container roottestconfigsubstitutions-gw3-node6-1 Stopped Stderr: Container roottestconfigsubstitutions-gw3-node6-1 Removing Stderr: Container roottestconfigsubstitutions-gw3-node2-1 Removed Stderr: Container roottestconfigsubstitutions-gw3-node1-1 Removed Stderr: Container roottestconfigsubstitutions-gw3-node5-1 Removed Stderr: Container roottestconfigsubstitutions-gw3-node3-1 Removed Stderr: Container roottestconfigsubstitutions-gw3-zoo3-1 Stopping Stderr: Container roottestconfigsubstitutions-gw3-node8-1 Removed Stderr: Container roottestconfigsubstitutions-gw3-zoo1-1 Stopping Stderr: Container roottestconfigsubstitutions-gw3-zoo2-1 Stopping Stderr: Container roottestconfigsubstitutions-gw3-zoo1-1 Stopped Stderr: Container roottestconfigsubstitutions-gw3-zoo1-1 Removing Stderr: Container roottestconfigsubstitutions-gw3-node6-1 Removed Stderr: Container roottestconfigsubstitutions-gw3-node7-1 Removed Stderr: Container roottestconfigsubstitutions-gw3-zoo3-1 Stopped Stderr: Container roottestconfigsubstitutions-gw3-zoo3-1 Removing Stderr: Container roottestconfigsubstitutions-gw3-zoo2-1 Stopped Stderr: Container roottestconfigsubstitutions-gw3-zoo2-1 Removing Stderr: Container roottestconfigsubstitutions-gw3-node4-1 Removed Stderr: Container roottestconfigsubstitutions-gw3-zoo1-1 Removed Stderr: Container roottestconfigsubstitutions-gw3-zoo3-1 Removed Stderr: Container roottestconfigsubstitutions-gw3-zoo2-1 Removed Stderr: Network roottestconfigsubstitutions-gw3_default Removing Stderr: Network roottestconfigsubstitutions-gw3_default Removed Cleanup called Executing query SELECT COUNT() FROM test_table_replicated on node1 Docker networks for project roottestconfigsubstitutions-gw3 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestconfigsubstitutions-gw3 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestconfigsubstitutions-gw3 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestconfigsubstitutions-gw3-.*-1$' --format '{{.ID}}:{{.Names}}'] Executing query SELECT COUNT() FROM test_table_replicated on node2 Unstopped containers: {} No running containers for project: roottestconfigsubstitutions-gw3 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:1 Volumes pruned: 1 Executing query SELECT COUNT() FROM test_table_replicated on node3 test_compatibility_merge_tree_settings/test.py::test_check_projections_compatibility Running tests in /ClickHouse/tests/integration/test_compatibility_merge_tree_settings/test.py Cluster start called. is_up=False Docker networks for project roottestcompatibilitymergetreesettings-gw3 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestcompatibilitymergetreesettings-gw3 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestcompatibilitymergetreesettings-gw3 are DRIVER VOLUME NAME Cleanup called Executing query SELECT COUNT() FROM test_table_replicated on node4 Docker networks for project roottestcompatibilitymergetreesettings-gw3 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestcompatibilitymergetreesettings-gw3 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestcompatibilitymergetreesettings-gw3 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestcompatibilitymergetreesettings-gw3-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestcompatibilitymergetreesettings-gw3 Trying to prune unused networks... Executing query TRUNCATE TABLE test_table_replicated on node1 Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:1 Volumes pruned: 1 Setup directory for instance: node1 Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files [] to /ClickHouse/tests/integration/test_compatibility_merge_tree_settings/_instances-0-gw3/node1/configs/config.d Setup database dir /ClickHouse/tests/integration/test_compatibility_merge_tree_settings/_instances-0-gw3/node1/database Setup logs dir /ClickHouse/tests/integration/test_compatibility_merge_tree_settings/_instances-0-gw3/node1/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Executing query TRUNCATE TABLE test_table_replicated on node2 Setup directory for instance: node2 Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files [] to /ClickHouse/tests/integration/test_compatibility_merge_tree_settings/_instances-0-gw3/node2/configs/config.d Setup database dir /ClickHouse/tests/integration/test_compatibility_merge_tree_settings/_instances-0-gw3/node2/database Setup logs dir /ClickHouse/tests/integration/test_compatibility_merge_tree_settings/_instances-0-gw3/node2/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Setup directory for instance: node3 Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_compatibility_merge_tree_settings/configs/mt_settings.xml'] to /ClickHouse/tests/integration/test_compatibility_merge_tree_settings/_instances-0-gw3/node3/configs/config.d Setup database dir /ClickHouse/tests/integration/test_compatibility_merge_tree_settings/_instances-0-gw3/node3/database Setup logs dir /ClickHouse/tests/integration/test_compatibility_merge_tree_settings/_instances-0-gw3/node3/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'keeper_binary': '/clickhouse', 'keeper_cmd_prefix': 'clickhouse keeper', 'image': 'altinityinfra/integration-test:8b2301119731', 'user': '0', 'keeper_fs': 'bind', 'keeper_logs_dir1': '/ClickHouse/tests/integration/test_compatibility_merge_tree_settings/_instances-0-gw3/keeper1/log', 'keeper_config_dir1': '/ClickHouse/tests/integration/test_compatibility_merge_tree_settings/_instances-0-gw3/keeper1/config', 'keeper_db_dir1': '/ClickHouse/tests/integration/test_compatibility_merge_tree_settings/_instances-0-gw3/keeper1/coordination', 'keeper_logs_dir2': '/ClickHouse/tests/integration/test_compatibility_merge_tree_settings/_instances-0-gw3/keeper2/log', 'keeper_config_dir2': '/ClickHouse/tests/integration/test_compatibility_merge_tree_settings/_instances-0-gw3/keeper2/config', 'keeper_db_dir2': '/ClickHouse/tests/integration/test_compatibility_merge_tree_settings/_instances-0-gw3/keeper2/coordination', 'keeper_logs_dir3': '/ClickHouse/tests/integration/test_compatibility_merge_tree_settings/_instances-0-gw3/keeper3/log', 'keeper_config_dir3': '/ClickHouse/tests/integration/test_compatibility_merge_tree_settings/_instances-0-gw3/keeper3/config', 'keeper_db_dir3': '/ClickHouse/tests/integration/test_compatibility_merge_tree_settings/_instances-0-gw3/keeper3/coordination'} stored in /ClickHouse/tests/integration/test_compatibility_merge_tree_settings/_instances-0-gw3/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_compatibility_merge_tree_settings/_instances-0-gw3/.env --project-name roottestcompatibilitymergetreesettings-gw3 --file /ClickHouse/tests/integration/test_compatibility_merge_tree_settings/_instances-0-gw3/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_compatibility_merge_tree_settings/_instances-0-gw3/node2/docker-compose.yml --file /ClickHouse/tests/integration/test_compatibility_merge_tree_settings/_instances-0-gw3/node3/docker-compose.yml pull] Executing query TRUNCATE TABLE test_table_replicated on node3 Executing query TRUNCATE TABLE test_table_replicated on node4 Command:[docker compose --env-file /ClickHouse/tests/integration/test_alter_on_mixed_type_cluster/_instances-0-gw7/.env --project-name roottestalteronmixedtypecluster-gw7 --file /ClickHouse/tests/integration/test_alter_on_mixed_type_cluster/_instances-0-gw7/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_alter_on_mixed_type_cluster/_instances-0-gw7/node2/docker-compose.yml --file /ClickHouse/tests/integration/test_alter_on_mixed_type_cluster/_instances-0-gw7/node3/docker-compose.yml --file /ClickHouse/tests/integration/test_alter_on_mixed_type_cluster/_instances-0-gw7/node4/docker-compose.yml stop --timeout 20] [gw7] PASSED test_alter_on_mixed_type_cluster/test.py::test_alter_replicated_on_cluster Executing query RESTORE TABLE tbl ON CLUSTER 'cluster' FROM Disk('backups', '10') on node1 Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw0] PASSED test_backup_restore_on_cluster/test.py::test_mutation Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 test_backup_restore_on_cluster/test.py::test_projection Executing query CREATE TABLE tbl ON CLUSTER 'cluster' (x UInt32, y String) ENGINE=ReplicatedMergeTree('/clickhouse/tables/tbl/', '{replica}') ORDER BY y PARTITION BY x%10 on node1 Executing query INSERT INTO tbl SELECT number, toString(number) FROM numbers(3) on node1 Executing query ALTER TABLE tbl ADD PROJECTION prjmax (SELECT MAX(x)) on node1 Executing query INSERT INTO tbl VALUES (100, 'a'), (101, 'b') on node1 Executing query SELECT count() FROM system.projection_parts WHERE database='default' AND table='tbl' AND name='prjmax' on node1 Executing query BACKUP TABLE tbl ON CLUSTER 'cluster' TO Disk('backups', '11') on node1 Executing query DROP TABLE tbl ON CLUSTER 'cluster' SYNC on node1 Executing query SELECT count() FROM system.projection_parts WHERE database='default' AND table='tbl' AND name='prjmax' on node1 Executing query RESTORE TABLE tbl FROM Disk('backups', '11') on node1 Executing query SELECT * FROM tbl ORDER BY x on node1 Executing query SELECT count() FROM system.projection_parts WHERE database='default' AND table='tbl' AND name='prjmax' on node1 Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw0] PASSED test_backup_restore_on_cluster/test.py::test_projection Executing query SELECT hostName() FROM clusterAllReplicas('two_shards_three_nodes', system.one) ORDER BY ALL settings skip_unavailable_shards=1 on node1 [gw1] PASSED test_cluster_all_replicas/test.py::test_skip_unavailable_replica[one_shard_three_nodes] test_cluster_all_replicas/test.py::test_skip_unavailable_replica[two_shards_three_nodes] Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Stderr: node2 Skipped - Image is already being pulled by node1 Stderr: node3 Skipped - Image is already being pulled by node1 Stderr: node1 Pulling Stderr: node1 Pulled ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_alternative_keeper_config/_instances-0-gw2/.env --project-name roottestalternativekeeperconfig-gw2 --file /ClickHouse/tests/integration/test_alternative_keeper_config/_instances-0-gw2/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_alternative_keeper_config/_instances-0-gw2/node2/docker-compose.yml --file /ClickHouse/tests/integration/test_alternative_keeper_config/_instances-0-gw2/node3/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_alternative_keeper_config/_instances-0-gw2/.env --project-name roottestalternativekeeperconfig-gw2 --file /ClickHouse/tests/integration/test_alternative_keeper_config/_instances-0-gw2/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_alternative_keeper_config/_instances-0-gw2/node2/docker-compose.yml --file /ClickHouse/tests/integration/test_alternative_keeper_config/_instances-0-gw2/node3/docker-compose.yml up -d --no-recreate] Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Stderr: Container roottestattachpartitionwithlargedestination-gw8-node-1 Stopping Stderr: Container roottestattachpartitionwithlargedestination-gw8-node-1 Stopped Stderr: Container roottestattachpartitionwithlargedestination-gw8-zoo1-1 Stopping Stderr: Container roottestattachpartitionwithlargedestination-gw8-zoo2-1 Stopping Stderr: Container roottestattachpartitionwithlargedestination-gw8-zoo3-1 Stopping Stderr: Container roottestattachpartitionwithlargedestination-gw8-zoo2-1 Stopped Stderr: Container roottestattachpartitionwithlargedestination-gw8-zoo1-1 Stopped Stderr: Container roottestattachpartitionwithlargedestination-gw8-zoo3-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_attach_partition_with_large_destination/_instances-0-gw8/node/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_attach_partition_with_large_destination/_instances-0-gw8/node/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_attach_partition_with_large_destination/_instances-0-gw8/.env --project-name roottestattachpartitionwithlargedestination-gw8 --file /ClickHouse/tests/integration/test_attach_partition_with_large_destination/_instances-0-gw8/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml down --volumes] Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Stderr: node3 Skipped - Image is already being pulled by node2 Stderr: node1 Skipped - Image is already being pulled by node2 Stderr: zoo2 Skipped - Image is already being pulled by node2 Stderr: zoo3 Skipped - Image is already being pulled by node2 Stderr: zoo1 Skipped - Image is already being pulled by node2 Stderr: node2 Pulling Stderr: node2 Pulled Setup ZooKeeper Creating internal ZooKeeper dirs: ['/ClickHouse/tests/integration/test_compatibility_merge_tree_settings/_instances-0-gw3/keeper1/log', '/ClickHouse/tests/integration/test_compatibility_merge_tree_settings/_instances-0-gw3/keeper1/config', '/ClickHouse/tests/integration/test_compatibility_merge_tree_settings/_instances-0-gw3/keeper1/coordination', '/ClickHouse/tests/integration/test_compatibility_merge_tree_settings/_instances-0-gw3/keeper2/log', '/ClickHouse/tests/integration/test_compatibility_merge_tree_settings/_instances-0-gw3/keeper2/config', '/ClickHouse/tests/integration/test_compatibility_merge_tree_settings/_instances-0-gw3/keeper2/coordination', '/ClickHouse/tests/integration/test_compatibility_merge_tree_settings/_instances-0-gw3/keeper3/log', '/ClickHouse/tests/integration/test_compatibility_merge_tree_settings/_instances-0-gw3/keeper3/config', '/ClickHouse/tests/integration/test_compatibility_merge_tree_settings/_instances-0-gw3/keeper3/coordination'] Command:[docker compose --project-name roottestcompatibilitymergetreesettings-gw3 --env-file /ClickHouse/tests/integration/test_compatibility_merge_tree_settings/_instances-0-gw3/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --verbose up -d] test_backup_restore_on_cluster/test.py::test_replicated_database Executing query CREATE DATABASE mydb ON CLUSTER 'cluster' ENGINE=Replicated('/clickhouse/path/','{shard}','{replica}') on node1 Executing query SYSTEM FLUSH LOGS on node1 Executing query CREATE TABLE mydb.tbl(x UInt8, y String) ENGINE=ReplicatedMergeTree ORDER BY x on node1 Executing query SELECT max(value) FROM ( SELECT toStartOfInterval(event_time, toIntervalSecond(1)) AS t, avg(value) AS value FROM system.asynchronous_metric_log WHERE event_time >= now() - 60 AND metric = 'OSUserTime' GROUP BY t ) SETTINGS max_threads = 1 on node1 Stderr: Container roottestattachpartitionwithlargedestination-gw8-node-1 Stopping Stderr: Container roottestattachpartitionwithlargedestination-gw8-node-1 Stopped Stderr: Container roottestattachpartitionwithlargedestination-gw8-node-1 Removing Stderr: Container roottestattachpartitionwithlargedestination-gw8-node-1 Removed Stderr: Container roottestattachpartitionwithlargedestination-gw8-zoo1-1 Stopping Stderr: Container roottestattachpartitionwithlargedestination-gw8-zoo2-1 Stopping Stderr: Container roottestattachpartitionwithlargedestination-gw8-zoo3-1 Stopping Stderr: Container roottestattachpartitionwithlargedestination-gw8-zoo1-1 Stopped Stderr: Container roottestattachpartitionwithlargedestination-gw8-zoo1-1 Removing Stderr: Container roottestattachpartitionwithlargedestination-gw8-zoo2-1 Stopped Stderr: Container roottestattachpartitionwithlargedestination-gw8-zoo2-1 Removing Stderr: Container roottestattachpartitionwithlargedestination-gw8-zoo3-1 Stopped Stderr: Container roottestattachpartitionwithlargedestination-gw8-zoo3-1 Removing Stderr: Container roottestattachpartitionwithlargedestination-gw8-zoo3-1 Removed Stderr: Container roottestattachpartitionwithlargedestination-gw8-zoo2-1 Removed Stderr: Container roottestattachpartitionwithlargedestination-gw8-zoo1-1 Removed Stderr: Network roottestattachpartitionwithlargedestination-gw8_default Removing Stderr: Network roottestattachpartitionwithlargedestination-gw8_default Removed Cleanup called Docker networks for project roottestattachpartitionwithlargedestination-gw8 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestattachpartitionwithlargedestination-gw8 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestattachpartitionwithlargedestination-gw8 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestattachpartitionwithlargedestination-gw8-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestattachpartitionwithlargedestination-gw8 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:1 Volumes pruned: 1 test_async_connect_to_multiple_ips/test.py::test Running tests in /ClickHouse/tests/integration/test_async_connect_to_multiple_ips/test.py Cluster start called. is_up=False Docker networks for project roottestasyncconnecttomultipleips-gw8 are NETWORK ID NAME DRIVER SCOPE Stderr: Network roottestalternativekeeperconfig-gw2_default Creating Stderr: Network roottestalternativekeeperconfig-gw2_default Created Stderr: Container roottestalternativekeeperconfig-gw2-node1-1 Creating Stderr: Container roottestalternativekeeperconfig-gw2-node2-1 Creating Stderr: Container roottestalternativekeeperconfig-gw2-node3-1 Creating Stderr: Container roottestalternativekeeperconfig-gw2-node1-1 Created Stderr: Container roottestalternativekeeperconfig-gw2-node3-1 Created Stderr: Container roottestalternativekeeperconfig-gw2-node2-1 Created Stderr: Container roottestalternativekeeperconfig-gw2-node1-1 Starting Stderr: Container roottestalternativekeeperconfig-gw2-node2-1 Starting Stderr: Container roottestalternativekeeperconfig-gw2-node3-1 Starting Stderr: Container roottestalternativekeeperconfig-gw2-node1-1 Started Stderr: Container roottestalternativekeeperconfig-gw2-node2-1 Started Stderr: Container roottestalternativekeeperconfig-gw2-node3-1 Started ClickHouse instance created get_instance_ip instance_name=node1 Docker containers for project roottestasyncconnecttomultipleips-gw8 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES http://localhost:None "GET /v1.46/containers/roottestalternativekeeperconfig-gw2-node1-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottestalternativekeeperconfig-gw2-node1-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node1, ip: 172.16.1.3... http://localhost:None "GET /v1.46/containers/roottestalternativekeeperconfig-gw2-node1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/ad5f015c0b2699513675a98652510fb4f126ba3b648decf08ff00c3af7abe204/json HTTP/1.1" 200 None Docker volumes for project roottestasyncconnecttomultipleips-gw8 are DRIVER VOLUME NAME Cleanup called Executing query SYSTEM FLUSH LOGS on node2 Docker networks for project roottestasyncconnecttomultipleips-gw8 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestasyncconnecttomultipleips-gw8 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestasyncconnecttomultipleips-gw8 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestasyncconnecttomultipleips-gw8-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestasyncconnecttomultipleips-gw8 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] http://localhost:None "GET /v1.46/containers/ad5f015c0b2699513675a98652510fb4f126ba3b648decf08ff00c3af7abe204/json HTTP/1.1" 200 None Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Executing query SYSTEM SYNC DATABASE REPLICA mydb on node2 Stdout:1 Volumes pruned: 1 Setup directory for instance: node1 Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_async_connect_to_multiple_ips/configs/listen_host.xml'] to /ClickHouse/tests/integration/test_async_connect_to_multiple_ips/_instances-0-gw8/node1/configs/config.d Setup database dir /ClickHouse/tests/integration/test_async_connect_to_multiple_ips/_instances-0-gw8/node1/database Setup logs dir /ClickHouse/tests/integration/test_async_connect_to_multiple_ips/_instances-0-gw8/node1/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Setup directory for instance: node2 Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_async_connect_to_multiple_ips/configs/listen_host.xml'] to /ClickHouse/tests/integration/test_async_connect_to_multiple_ips/_instances-0-gw8/node2/configs/config.d Setup database dir /ClickHouse/tests/integration/test_async_connect_to_multiple_ips/_instances-0-gw8/node2/database Setup logs dir /ClickHouse/tests/integration/test_async_connect_to_multiple_ips/_instances-0-gw8/node2/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'keeper_binary': '/clickhouse', 'keeper_cmd_prefix': 'clickhouse keeper', 'image': 'altinityinfra/integration-test:8b2301119731', 'user': '0', 'keeper_fs': 'bind', 'keeper_logs_dir1': '/ClickHouse/tests/integration/test_async_connect_to_multiple_ips/_instances-0-gw8/keeper1/log', 'keeper_config_dir1': '/ClickHouse/tests/integration/test_async_connect_to_multiple_ips/_instances-0-gw8/keeper1/config', 'keeper_db_dir1': '/ClickHouse/tests/integration/test_async_connect_to_multiple_ips/_instances-0-gw8/keeper1/coordination', 'keeper_logs_dir2': '/ClickHouse/tests/integration/test_async_connect_to_multiple_ips/_instances-0-gw8/keeper2/log', 'keeper_config_dir2': '/ClickHouse/tests/integration/test_async_connect_to_multiple_ips/_instances-0-gw8/keeper2/config', 'keeper_db_dir2': '/ClickHouse/tests/integration/test_async_connect_to_multiple_ips/_instances-0-gw8/keeper2/coordination', 'keeper_logs_dir3': '/ClickHouse/tests/integration/test_async_connect_to_multiple_ips/_instances-0-gw8/keeper3/log', 'keeper_config_dir3': '/ClickHouse/tests/integration/test_async_connect_to_multiple_ips/_instances-0-gw8/keeper3/config', 'keeper_db_dir3': '/ClickHouse/tests/integration/test_async_connect_to_multiple_ips/_instances-0-gw8/keeper3/coordination'} stored in /ClickHouse/tests/integration/test_async_connect_to_multiple_ips/_instances-0-gw8/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Executing query SELECT max(value) FROM ( SELECT toStartOfInterval(event_time, toIntervalSecond(1)) AS t, avg(value) AS value FROM system.asynchronous_metric_log WHERE event_time >= now() - 60 AND metric = 'OSUserTime' GROUP BY t ) SETTINGS max_threads = 1 on node2 http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_async_connect_to_multiple_ips/_instances-0-gw8/.env --project-name roottestasyncconnecttomultipleips-gw8 --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_net.yml --file /ClickHouse/tests/integration/test_async_connect_to_multiple_ips/_instances-0-gw8/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_async_connect_to_multiple_ips/_instances-0-gw8/node2/docker-compose.yml pull] Executing query INSERT INTO mydb.tbl VALUES (1, 'Don''t') on node1 http://localhost:None "GET /v1.46/containers/ad5f015c0b2699513675a98652510fb4f126ba3b648decf08ff00c3af7abe204/json HTTP/1.1" 200 None Executing query SYSTEM FLUSH LOGS on node1 Executing query INSERT INTO mydb.tbl VALUES (2, 'count') on node2 http://localhost:None "GET /v1.46/containers/ad5f015c0b2699513675a98652510fb4f126ba3b648decf08ff00c3af7abe204/json HTTP/1.1" 200 None Executing query INSERT INTO mydb.tbl VALUES (3, 'your') on node1 Executing query SELECT max(value) FROM ( SELECT toStartOfInterval(event_time, toIntervalSecond(1)) AS t, avg(value) AS value FROM system.asynchronous_metric_log WHERE event_time >= now() - 60 AND metric = 'OSIdleTime' GROUP BY t ) SETTINGS max_threads = 1 on node1 Executing query INSERT INTO mydb.tbl VALUES (4, 'chickens') on node2 Executing query SYSTEM FLUSH LOGS on node2 http://localhost:None "GET /v1.46/containers/ad5f015c0b2699513675a98652510fb4f126ba3b648decf08ff00c3af7abe204/json HTTP/1.1" 200 None Executing query SELECT max(value) FROM ( SELECT toStartOfInterval(event_time, toIntervalSecond(1)) AS t, avg(value) AS value FROM system.asynchronous_metric_log WHERE event_time >= now() - 60 AND metric = 'OSIdleTime' GROUP BY t ) SETTINGS max_threads = 1 on node2 Executing query OPTIMIZE TABLE mydb.tbl ON CLUSTER 'cluster' FINAL on node1 http://localhost:None "GET /v1.46/containers/ad5f015c0b2699513675a98652510fb4f126ba3b648decf08ff00c3af7abe204/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/ad5f015c0b2699513675a98652510fb4f126ba3b648decf08ff00c3af7abe204/json HTTP/1.1" 200 None Stderr:time="2025-04-02T03:34:20Z" level=trace msg="Docker Desktop integration not enabled" Stderr: Network roottestcompatibilitymergetreesettings-gw3_default Creating Stderr: Network roottestcompatibilitymergetreesettings-gw3_default Created Stderr: Container roottestcompatibilitymergetreesettings-gw3-zoo1-1 Creating Stderr: Container roottestcompatibilitymergetreesettings-gw3-zoo2-1 Creating Stderr: Container roottestcompatibilitymergetreesettings-gw3-zoo3-1 Creating Stderr: Container roottestcompatibilitymergetreesettings-gw3-zoo3-1 Created Stderr: Container roottestcompatibilitymergetreesettings-gw3-zoo1-1 Created Stderr: Container roottestcompatibilitymergetreesettings-gw3-zoo2-1 Created Stderr: Container roottestcompatibilitymergetreesettings-gw3-zoo1-1 Starting Stderr: Container roottestcompatibilitymergetreesettings-gw3-zoo3-1 Starting Stderr: Container roottestcompatibilitymergetreesettings-gw3-zoo2-1 Starting Stderr: Container roottestcompatibilitymergetreesettings-gw3-zoo1-1 Started Stderr: Container roottestcompatibilitymergetreesettings-gw3-zoo2-1 Started Stderr: Container roottestcompatibilitymergetreesettings-gw3-zoo3-1 Started Stderr:time="2025-04-02T03:34:21Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T03:34:21Z" level=debug msg="otel error" error="" Wait ZooKeeper to start get_instance_ip instance_name=zoo1 http://localhost:None "GET /v1.46/containers/roottestcompatibilitymergetreesettings-gw3-zoo1-1/json HTTP/1.1" 200 None get_kazoo_client: zoo1, ip:172.16.2.2, port:2181, use_ssl:False Connecting to 172.16.2.2(172.16.2.2):2181, use_ssl: False Connection dropped: socket connection error: Connection refused http://localhost:None "GET /v1.46/containers/ad5f015c0b2699513675a98652510fb4f126ba3b648decf08ff00c3af7abe204/json HTTP/1.1" 200 None Command:[docker compose --env-file /ClickHouse/tests/integration/test_async_metrics_in_cgroup/_instances-0-gw6/.env --project-name roottestasyncmetricsincgroup-gw6 --file /ClickHouse/tests/integration/test_async_metrics_in_cgroup/_instances-0-gw6/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_async_metrics_in_cgroup/_instances-0-gw6/node2/docker-compose.yml stop --timeout 20] [gw6] PASSED test_async_metrics_in_cgroup/test.py::test_user_cpu_accounting Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster' mydb.tbl on node1 http://localhost:None "GET /v1.46/containers/ad5f015c0b2699513675a98652510fb4f126ba3b648decf08ff00c3af7abe204/json HTTP/1.1" 200 None Connecting to 172.16.2.2(172.16.2.2):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Executing query SELECT * FROM mydb.tbl ORDER BY x on node1 http://localhost:None "GET /v1.46/containers/ad5f015c0b2699513675a98652510fb4f126ba3b648decf08ff00c3af7abe204/json HTTP/1.1" 200 None Executing query SELECT * FROM mydb.tbl ORDER BY x on node2 Connecting to 172.16.2.2(172.16.2.2):2181, use_ssl: False Connection dropped: socket connection error: Connection refused http://localhost:None "GET /v1.46/containers/ad5f015c0b2699513675a98652510fb4f126ba3b648decf08ff00c3af7abe204/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/ad5f015c0b2699513675a98652510fb4f126ba3b648decf08ff00c3af7abe204/json HTTP/1.1" 200 None Executing query BACKUP DATABASE mydb ON CLUSTER 'cluster' TO Disk('backups', '12') SETTINGS replica_num=2 on node1 http://localhost:None "GET /v1.46/containers/ad5f015c0b2699513675a98652510fb4f126ba3b648decf08ff00c3af7abe204/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/ad5f015c0b2699513675a98652510fb4f126ba3b648decf08ff00c3af7abe204/json HTTP/1.1" 200 None Stderr: Container roottestasyncmetricsincgroup-gw6-node1-1 Stopping Stderr: Container roottestasyncmetricsincgroup-gw6-node2-1 Stopping Stderr: Container roottestasyncmetricsincgroup-gw6-node2-1 Stopped Stderr: Container roottestasyncmetricsincgroup-gw6-node1-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_async_metrics_in_cgroup/_instances-0-gw6/node1/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_async_metrics_in_cgroup/_instances-0-gw6/node1/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_async_metrics_in_cgroup/_instances-0-gw6/node2/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_async_metrics_in_cgroup/_instances-0-gw6/node2/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_async_metrics_in_cgroup/_instances-0-gw6/.env --project-name roottestasyncmetricsincgroup-gw6 --file /ClickHouse/tests/integration/test_async_metrics_in_cgroup/_instances-0-gw6/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_async_metrics_in_cgroup/_instances-0-gw6/node2/docker-compose.yml down --volumes] Executing query DROP DATABASE mydb ON CLUSTER 'cluster' SYNC on node1 http://localhost:None "GET /v1.46/containers/ad5f015c0b2699513675a98652510fb4f126ba3b648decf08ff00c3af7abe204/json HTTP/1.1" 200 None Executing query RESTORE DATABASE mydb ON CLUSTER 'cluster' FROM Disk('backups', '12') on node1 Connecting to 172.16.2.2(172.16.2.2):2181, use_ssl: False Connection dropped: socket connection error: Connection refused http://localhost:None "GET /v1.46/containers/ad5f015c0b2699513675a98652510fb4f126ba3b648decf08ff00c3af7abe204/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/ad5f015c0b2699513675a98652510fb4f126ba3b648decf08ff00c3af7abe204/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/ad5f015c0b2699513675a98652510fb4f126ba3b648decf08ff00c3af7abe204/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/ad5f015c0b2699513675a98652510fb4f126ba3b648decf08ff00c3af7abe204/json HTTP/1.1" 200 None Connection dropped: socket connection error: Connection refused Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster' mydb.tbl on node1 http://localhost:None "GET /v1.46/containers/ad5f015c0b2699513675a98652510fb4f126ba3b648decf08ff00c3af7abe204/json HTTP/1.1" 200 None Executing query SELECT * FROM mydb.tbl ORDER BY x on node1 Executing query SELECT * FROM mydb.tbl ORDER BY x on node2 Stderr: Container roottestasyncmetricsincgroup-gw6-node2-1 Stopping Stderr: Container roottestasyncmetricsincgroup-gw6-node1-1 Stopping Stderr: Container roottestasyncmetricsincgroup-gw6-node2-1 Stopped Stderr: Container roottestasyncmetricsincgroup-gw6-node2-1 Removing Stderr: Container roottestasyncmetricsincgroup-gw6-node1-1 Stopped Stderr: Container roottestasyncmetricsincgroup-gw6-node1-1 Removing Stderr: Container roottestasyncmetricsincgroup-gw6-node2-1 Removed Stderr: Container roottestasyncmetricsincgroup-gw6-node1-1 Removed Stderr: Network roottestasyncmetricsincgroup-gw6_default Removing Stderr: Network roottestasyncmetricsincgroup-gw6_default Removed Cleanup called http://localhost:None "GET /v1.46/containers/ad5f015c0b2699513675a98652510fb4f126ba3b648decf08ff00c3af7abe204/json HTTP/1.1" 200 None Docker networks for project roottestasyncmetricsincgroup-gw6 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestasyncmetricsincgroup-gw6 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestasyncmetricsincgroup-gw6 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestasyncmetricsincgroup-gw6-.*-1$' --format '{{.ID}}:{{.Names}}'] Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw0] PASSED test_backup_restore_on_cluster/test.py::test_replicated_database Unstopped containers: {} No running containers for project: roottestasyncmetricsincgroup-gw6 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] http://localhost:None "GET /v1.46/containers/ad5f015c0b2699513675a98652510fb4f126ba3b648decf08ff00c3af7abe204/json HTTP/1.1" 200 None Stdout:1 Volumes pruned: 1 test_aggregation_memory_efficient/test.py::test_remote Running tests in /ClickHouse/tests/integration/test_aggregation_memory_efficient/test.py Cluster start called. is_up=False Docker networks for project roottestaggregationmemoryefficient-gw6 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestaggregationmemoryefficient-gw6 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestaggregationmemoryefficient-gw6 are DRIVER VOLUME NAME Cleanup called Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Docker networks for project roottestaggregationmemoryefficient-gw6 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestaggregationmemoryefficient-gw6 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES http://localhost:None "GET /v1.46/containers/ad5f015c0b2699513675a98652510fb4f126ba3b648decf08ff00c3af7abe204/json HTTP/1.1" 200 None ClickHouse node1 started get_instance_ip instance_name=node2 http://localhost:None "GET /v1.46/containers/roottestalternativekeeperconfig-gw2-node2-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node2 http://localhost:None "GET /v1.46/containers/roottestalternativekeeperconfig-gw2-node2-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node2, ip: 172.16.1.2... http://localhost:None "GET /v1.46/containers/roottestalternativekeeperconfig-gw2-node2-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/820eb54910beae1983df52aa5b9a892e03216a08f182640c7abb1d87b7c16573/json HTTP/1.1" 200 None ClickHouse node2 started get_instance_ip instance_name=node3 http://localhost:None "GET /v1.46/containers/roottestalternativekeeperconfig-gw2-node3-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node3 http://localhost:None "GET /v1.46/containers/roottestalternativekeeperconfig-gw2-node3-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node3, ip: 172.16.1.4... http://localhost:None "GET /v1.46/containers/roottestalternativekeeperconfig-gw2-node3-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/e79923898f6bceb7828395a5034b15b85c0171d2fc9d03b3c7f267aa75472b15/json HTTP/1.1" 200 None ClickHouse node3 started Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'test_cluster' SYNC on node1 Docker volumes for project roottestaggregationmemoryefficient-gw6 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestaggregationmemoryefficient-gw6-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestaggregationmemoryefficient-gw6 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Stdout:1 Volumes pruned: 1 Setup directory for instance: node1 Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files [] to /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw6/node1/configs/config.d Setup database dir /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw6/node1/database Setup logs dir /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw6/node1/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Setup directory for instance: node2 Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files [] to /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw6/node2/configs/config.d Setup database dir /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw6/node2/database Setup logs dir /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw6/node2/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw'} stored in /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw6/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw6/.env --project-name roottestaggregationmemoryefficient-gw6 --file /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw6/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw6/node2/docker-compose.yml pull] Connection dropped: socket connection error: Connection refused Executing query CREATE TABLE tbl ON CLUSTER 'test_cluster' ( id Int64, str String ) ENGINE=ReplicatedMergeTree('/clickhouse/tables/tbl/', '{replica}') ORDER BY id on node1 Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Connecting to 172.16.2.3(172.16.2.3):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Executing query INSERT INTO tbl VALUES (1, 'str1') on node1 Executing query INSERT INTO tbl VALUES (1, 'str1') on node2 test_backup_restore_on_cluster/test.py::test_replicated_database_async Executing query CREATE DATABASE mydb ON CLUSTER 'cluster' ENGINE=Replicated('/clickhouse/path/','{shard}','{replica}') on node1 Connecting to 172.16.2.2(172.16.2.2):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Executing query INSERT INTO tbl VALUES (2, 'str2') on node3 Executing query CREATE TABLE mydb.tbl(x UInt8) ENGINE=ReplicatedMergeTree ORDER BY x on node1 Executing query SYSTEM SYNC REPLICA ON CLUSTER 'test_cluster' tbl on node1 Executing query SELECT * FROM tbl ORDER BY id on node1 Executing query CREATE TABLE mydb.tbl2(y String) ENGINE=ReplicatedMergeTree ORDER BY y on node1 Executing query CHECK TABLE tbl on node1 Executing query SELECT * FROM tbl ORDER BY id on node2 Executing query SYSTEM SYNC DATABASE REPLICA mydb on node2 Executing query CHECK TABLE tbl on node2 Executing query INSERT INTO mydb.tbl VALUES (1) on node1 Executing query SELECT * FROM tbl ORDER BY id on node3 Executing query INSERT INTO mydb.tbl VALUES (22) on node1 Executing query INSERT INTO mydb.tbl2 VALUES ('a') on node2 Executing query CHECK TABLE tbl on node3 Connecting to 172.16.2.3(172.16.2.3):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=38, time_out=30000, session_id=7, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Executing query INSERT INTO mydb.tbl2 VALUES ('bb') on node2 Command:[docker compose --env-file /ClickHouse/tests/integration/test_alternative_keeper_config/_instances-0-gw2/.env --project-name roottestalternativekeeperconfig-gw2 --file /ClickHouse/tests/integration/test_alternative_keeper_config/_instances-0-gw2/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_alternative_keeper_config/_instances-0-gw2/node2/docker-compose.yml --file /ClickHouse/tests/integration/test_alternative_keeper_config/_instances-0-gw2/node3/docker-compose.yml stop --timeout 20] [gw2] PASSED test_alternative_keeper_config/test.py::test_create_insert Executing query OPTIMIZE TABLE mydb.tbl ON CLUSTER 'cluster' FINAL on node1 Executing query OPTIMIZE TABLE mydb.tbl2 ON CLUSTER 'cluster' FINAL on node1 Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster' mydb.tbl on node1 Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster' mydb.tbl2 on node1 Connecting to 172.16.2.3(172.16.2.3):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=10, time_out=30000, session_id=4, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Executing query BACKUP DATABASE mydb ON CLUSTER 'cluster' TO Disk('backups', '13') ASYNC on node1 Executing query SELECT status, error FROM system.backups WHERE id='0ef5e1c0-4e9f-4b96-8111-8d326add3a8e' on node1 Executing query SELECT status, error FROM system.backups WHERE id='0ef5e1c0-4e9f-4b96-8111-8d326add3a8e' on node1 Executing query DROP DATABASE mydb ON CLUSTER 'cluster' SYNC on node1 Executing query RESTORE DATABASE mydb ON CLUSTER 'cluster' FROM Disk('backups', '13') ASYNC on node1 Executing query SELECT status, error FROM system.backups WHERE id='5c6873f9-601e-4699-8235-c1bd3f21f1b4' on node1 Connecting to 172.16.2.2(172.16.2.2):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost Stderr: upstream Pulling Stderr: backward Pulling Stderr: old_node Pulling Stderr: upstream Pulled Stderr: upstream_node Pulling Stderr: 17d0386c2fff Pulling fs layer Stderr: upstream_node Pulled Stderr: c225dbfce283 Pulling fs layer Stderr: 17d0386c2fff Pulling fs layer Stderr: c225dbfce283 Pulling fs layer Stderr: 2130ef613416 Pulling fs layer Stderr: 2130ef613416 Pulling fs layer Stderr: 4f4fb700ef54 Pulling fs layer Stderr: 4f4fb700ef54 Pulling fs layer Stderr: 6a3fe6bcecc5 Pulling fs layer Stderr: 6a3fe6bcecc5 Pulling fs layer Stderr: 33bc578e5e14 Pulling fs layer Stderr: 6d0df3aa878d Pulling fs layer Stderr: 33bc578e5e14 Pulling fs layer Stderr: 690edf6c102b Pulling fs layer Stderr: 87e28718f844 Pulling fs layer Stderr: 6d0df3aa878d Pulling fs layer Stderr: 33bc578e5e14 Waiting Stderr: 690edf6c102b Pulling fs layer Stderr: 690edf6c102b Waiting Stderr: 87e28718f844 Pulling fs layer Stderr: 6d0df3aa878d Waiting Stderr: 4f4fb700ef54 Waiting Stderr: 17d0386c2fff Downloading [> ] 298kB/27.51MB Stderr: 87e28718f844 Waiting Stderr: 6a3fe6bcecc5 Waiting Stderr: c225dbfce283 Downloading [> ] 100.7kB/8.787MB Stderr: 17d0386c2fff Downloading [> ] 298kB/27.51MB Stderr: 2130ef613416 Downloading [> ] 536.5kB/258.8MB Stderr: c225dbfce283 Downloading [> ] 100.7kB/8.787MB Stderr: 2130ef613416 Downloading [> ] 536.5kB/258.8MB Stderr: 4f4fb700ef54 Waiting Stderr: c225dbfce283 Verifying Checksum Stderr: c225dbfce283 Download complete Stderr: 6a3fe6bcecc5 Waiting Stderr: 4f4fb700ef54 Verifying Checksum Stderr: 33bc578e5e14 Waiting Stderr: 4f4fb700ef54 Download complete Stderr: 17d0386c2fff Download complete Stderr: 6d0df3aa878d Waiting Stderr: 6a3fe6bcecc5 Downloading [========> ] 3.646kB/22.46kB Stderr: 690edf6c102b Waiting Stderr: 6a3fe6bcecc5 Downloading [==================================================>] 22.46kB/22.46kB Stderr: 6a3fe6bcecc5 Verifying Checksum Stderr: 87e28718f844 Waiting Stderr: 6a3fe6bcecc5 Download complete Stderr: c225dbfce283 Verifying Checksum Stderr: 6d0df3aa878d Downloading [==================================================>] 116B/116B Stderr: c225dbfce283 Download complete Stderr: 6d0df3aa878d Verifying Checksum Stderr: 6d0df3aa878d Download complete Stderr: 4f4fb700ef54 Downloading [==================================================>] 32B/32B Stderr: 33bc578e5e14 Downloading [> ] 13.78kB/863.5kB Stderr: 4f4fb700ef54 Verifying Checksum Stderr: 33bc578e5e14 Downloading [==================================================>] 863.5kB/863.5kB Stderr: 33bc578e5e14 Verifying Checksum Stderr: 33bc578e5e14 Download complete Stderr: 4f4fb700ef54 Download complete Stderr: 690edf6c102b Downloading [==================================================>] 364B/364B Stderr: 17d0386c2fff Downloading [================================================> ] 26.53MB/27.51MB Stderr: 690edf6c102b Verifying Checksum Stderr: 690edf6c102b Download complete Stderr: 17d0386c2fff Verifying Checksum Stderr: 87e28718f844 Downloading [==================================================>] 2.936kB/2.936kB Stderr: 87e28718f844 Verifying Checksum Stderr: 17d0386c2fff Download complete Stderr: 87e28718f844 Download complete Stderr: 6a3fe6bcecc5 Downloading [========> ] 3.646kB/22.46kB Stderr: 2130ef613416 Downloading [====> ] 20.99MB/258.8MB Stderr: 6a3fe6bcecc5 Downloading [==================================================>] 22.46kB/22.46kB Stderr: 17d0386c2fff Extracting [> ] 294.9kB/27.51MB Stderr: 2130ef613416 Downloading [==========> ] 54.86MB/258.8MB Stderr: 6a3fe6bcecc5 Verifying Checksum Stderr: 17d0386c2fff Extracting [======> ] 3.834MB/27.51MB Stderr: 6a3fe6bcecc5 Download complete Stderr: 2130ef613416 Downloading [================> ] 84.83MB/258.8MB Stderr: 6d0df3aa878d Downloading [==================================================>] 116B/116B Stderr: 6d0df3aa878d Verifying Checksum Stderr: 17d0386c2fff Extracting [===============> ] 8.552MB/27.51MB Stderr: 6d0df3aa878d Download complete Stderr: 2130ef613416 Downloading [======================> ] 118.2MB/258.8MB Stderr: 33bc578e5e14 Downloading [> ] 13.78kB/863.5kB Stderr: 17d0386c2fff Extracting [===================> ] 10.62MB/27.51MB Stderr: 33bc578e5e14 Downloading [==================================================>] 863.5kB/863.5kB Stderr: 2130ef613416 Downloading [=============================> ] 152.1MB/258.8MB Stderr: 33bc578e5e14 Verifying Checksum Stderr: 17d0386c2fff Extracting [===========================> ] 15.34MB/27.51MB Stderr: 2130ef613416 Downloading [===================================> ] 185.3MB/258.8MB Stderr: 33bc578e5e14 Download complete Stderr: 690edf6c102b Downloading [==================================================>] 364B/364B Stderr: 17d0386c2fff Extracting [========================================> ] 22.41MB/27.51MB Stderr: 690edf6c102b Verifying Checksum Stderr: 2130ef613416 Downloading [==========================================> ] 218.6MB/258.8MB Stderr: 690edf6c102b Download complete Stderr: 2130ef613416 Downloading [===============================================> ] 248.1MB/258.8MB Stderr: 17d0386c2fff Extracting [==========================================> ] 23.3MB/27.51MB Stderr: 87e28718f844 Downloading [==================================================>] 2.936kB/2.936kB Stderr: 2130ef613416 Verifying Checksum Stderr: 2130ef613416 Download complete Stderr: 87e28718f844 Verifying Checksum Stderr: 17d0386c2fff Extracting [==============================================> ] 25.66MB/27.51MB Stderr: 17d0386c2fff Extracting [================================================> ] 26.54MB/27.51MB Stderr: 87e28718f844 Download complete Stderr: 17d0386c2fff Extracting [=================================================> ] 27.43MB/27.51MB Stderr: 17d0386c2fff Extracting [==================================================>] 27.51MB/27.51MB Stderr: 2130ef613416 Downloading [====> ] 20.99MB/258.8MB Stderr: 17d0386c2fff Pull complete Stderr: c225dbfce283 Extracting [> ] 98.3kB/8.787MB Stderr: 17d0386c2fff Extracting [> ] 294.9kB/27.51MB Stderr: c225dbfce283 Extracting [===============> ] 2.654MB/8.787MB Stderr: c225dbfce283 Extracting [===============================> ] 5.603MB/8.787MB Stderr: 2130ef613416 Downloading [==========> ] 54.86MB/258.8MB Stderr: c225dbfce283 Extracting [============================================> ] 7.766MB/8.787MB Stderr: c225dbfce283 Extracting [==============================================> ] 8.159MB/8.787MB Stderr: 17d0386c2fff Extracting [======> ] 3.834MB/27.51MB Stderr: c225dbfce283 Extracting [===============================================> ] 8.356MB/8.787MB Stderr: 2130ef613416 Downloading [================> ] 84.83MB/258.8MB Stderr: c225dbfce283 Extracting [================================================> ] 8.552MB/8.787MB Stderr: c225dbfce283 Extracting [==================================================>] 8.787MB/8.787MB Stderr: 17d0386c2fff Extracting [===============> ] 8.552MB/27.51MB Stderr: c225dbfce283 Pull complete Stderr: 2130ef613416 Downloading [======================> ] 118.2MB/258.8MB Stderr: 2130ef613416 Extracting [> ] 557.1kB/258.8MB Stderr: 2130ef613416 Extracting [> ] 5.014MB/258.8MB Stderr: 17d0386c2fff Extracting [===================> ] 10.62MB/27.51MB Stderr: 2130ef613416 Extracting [==> ] 12.26MB/258.8MB Stderr: 2130ef613416 Downloading [=============================> ] 152.1MB/258.8MB Stderr: 2130ef613416 Extracting [===> ] 20.05MB/258.8MB Stderr: 17d0386c2fff Extracting [===========================> ] 15.34MB/27.51MB Stderr: 2130ef613416 Extracting [=====> ] 26.18MB/258.8MB Stderr: 2130ef613416 Downloading [===================================> ] 185.3MB/258.8MB Stderr: 2130ef613416 Extracting [======> ] 35.65MB/258.8MB Stderr: 17d0386c2fff Extracting [========================================> ] 22.41MB/27.51MB Stderr: 2130ef613416 Extracting [========> ] 42.89MB/258.8MB Stderr: 2130ef613416 Extracting [=========> ] 49.02MB/258.8MB Stderr: 2130ef613416 Downloading [==========================================> ] 218.6MB/258.8MB Stderr: 2130ef613416 Extracting [==========> ] 55.15MB/258.8MB Stderr: 2130ef613416 Downloading [===============================================> ] 248.1MB/258.8MB Stderr: 2130ef613416 Extracting [===========> ] 59.6MB/258.8MB Stderr: 2130ef613416 Extracting [============> ] 65.18MB/258.8MB Stderr: 17d0386c2fff Extracting [==========================================> ] 23.3MB/27.51MB Stderr: 2130ef613416 Extracting [=============> ] 70.75MB/258.8MB Stderr: 2130ef613416 Verifying Checksum Stderr: 2130ef613416 Extracting [==============> ] 75.76MB/258.8MB Stderr: 2130ef613416 Extracting [===============> ] 81.89MB/258.8MB Stderr: 2130ef613416 Download complete Stderr: 2130ef613416 Extracting [=================> ] 90.24MB/258.8MB Stderr: 17d0386c2fff Extracting [==============================================> ] 25.66MB/27.51MB Stderr: 17d0386c2fff Extracting [================================================> ] 26.54MB/27.51MB Stderr: 2130ef613416 Extracting [===================> ] 100.3MB/258.8MB Stderr: 2130ef613416 Extracting [=====================> ] 110.3MB/258.8MB Stderr: 17d0386c2fff Extracting [=================================================> ] 27.43MB/27.51MB Stderr: 2130ef613416 Extracting [=======================> ] 120.3MB/258.8MB Stderr: 17d0386c2fff Extracting [==================================================>] 27.51MB/27.51MB Stderr: 2130ef613416 Extracting [========================> ] 128.7MB/258.8MB Stderr: 2130ef613416 Extracting [==========================> ] 137.6MB/258.8MB Stderr: 17d0386c2fff Pull complete Stderr: 2130ef613416 Extracting [============================> ] 148.2MB/258.8MB Stderr: 2130ef613416 Extracting [==============================> ] 157.1MB/258.8MB Stderr: c225dbfce283 Extracting [> ] 98.3kB/8.787MB Stderr: 2130ef613416 Extracting [===============================> ] 165.4MB/258.8MB Stderr: c225dbfce283 Extracting [===============> ] 2.654MB/8.787MB Stderr: 2130ef613416 Extracting [=================================> ] 172.1MB/258.8MB Stderr: c225dbfce283 Extracting [===============================> ] 5.603MB/8.787MB Stderr: 2130ef613416 Extracting [==================================> ] 181MB/258.8MB Stderr: c225dbfce283 Extracting [============================================> ] 7.766MB/8.787MB Stderr: 2130ef613416 Extracting [====================================> ] 191.1MB/258.8MB Stderr: 2130ef613416 Extracting [======================================> ] 200.5MB/258.8MB Stderr: c225dbfce283 Extracting [==============================================> ] 8.159MB/8.787MB Stderr: 2130ef613416 Extracting [=========================================> ] 212.2MB/258.8MB Stderr: 2130ef613416 Extracting [==========================================> ] 221.7MB/258.8MB Stderr: c225dbfce283 Extracting [===============================================> ] 8.356MB/8.787MB Stderr: 2130ef613416 Extracting [============================================> ] 229.5MB/258.8MB Stderr: 2130ef613416 Extracting [==============================================> ] 239.5MB/258.8MB Stderr: c225dbfce283 Extracting [================================================> ] 8.552MB/8.787MB Stderr: 2130ef613416 Extracting [================================================> ] 249.6MB/258.8MB Stderr: c225dbfce283 Extracting [==================================================>] 8.787MB/8.787MB Stderr: 2130ef613416 Extracting [=================================================> ] 257.4MB/258.8MB Stderr: c225dbfce283 Pull complete Stderr: 2130ef613416 Extracting [==================================================>] 258.8MB/258.8MB Stderr: 2130ef613416 Extracting [> ] 557.1kB/258.8MB Stderr: 2130ef613416 Pull complete Stderr: 2130ef613416 Extracting [> ] 5.014MB/258.8MB Stderr: 4f4fb700ef54 Extracting [==================================================>] 32B/32B Stderr: 2130ef613416 Extracting [==> ] 12.26MB/258.8MB Stderr: 4f4fb700ef54 Extracting [==================================================>] 32B/32B Stderr: 4f4fb700ef54 Pull complete Stderr: 2130ef613416 Extracting [===> ] 20.05MB/258.8MB Stderr: 6a3fe6bcecc5 Extracting [==================================================>] 22.46kB/22.46kB Stderr: 2130ef613416 Extracting [=====> ] 26.18MB/258.8MB Stderr: 6a3fe6bcecc5 Extracting [==================================================>] 22.46kB/22.46kB Stderr: 2130ef613416 Extracting [======> ] 35.65MB/258.8MB Stderr: 6a3fe6bcecc5 Pull complete Stderr: 33bc578e5e14 Extracting [=> ] 32.77kB/863.5kB Stderr: 33bc578e5e14 Extracting [==================================================>] 863.5kB/863.5kB Stderr: 33bc578e5e14 Extracting [==================================================>] 863.5kB/863.5kB Stderr: 2130ef613416 Extracting [========> ] 42.89MB/258.8MB Stderr: 33bc578e5e14 Pull complete Stderr: 6d0df3aa878d Extracting [==================================================>] 116B/116B Stderr: 2130ef613416 Extracting [=========> ] 49.02MB/258.8MB Stderr: 6d0df3aa878d Extracting [==================================================>] 116B/116B Stderr: 2130ef613416 Extracting [==========> ] 55.15MB/258.8MB Stderr: 6d0df3aa878d Pull complete Stderr: 2130ef613416 Extracting [===========> ] 59.6MB/258.8MB Stderr: 690edf6c102b Extracting [==================================================>] 364B/364B Stderr: 690edf6c102b Extracting [==================================================>] 364B/364B Stderr: 2130ef613416 Extracting [============> ] 65.18MB/258.8MB Stderr: 690edf6c102b Pull complete Stderr: 2130ef613416 Extracting [=============> ] 70.75MB/258.8MB Stderr: 2130ef613416 Extracting [==============> ] 75.76MB/258.8MB Stderr: 87e28718f844 Extracting [==================================================>] 2.936kB/2.936kB Stderr: 87e28718f844 Extracting [==================================================>] 2.936kB/2.936kB Stderr: 2130ef613416 Extracting [===============> ] 81.89MB/258.8MB Stderr: 87e28718f844 Pull complete Stderr: 2130ef613416 Extracting [=================> ] 90.24MB/258.8MB Stderr: old_node Pulled Stderr: 2130ef613416 Extracting [===================> ] 100.3MB/258.8MB ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_backward_compatibility/_instances-insert_profile_events-0-gw5/.env --project-name roottestbackwardcompatibilityinsertprofileevents-gw5 --file /ClickHouse/tests/integration/test_backward_compatibility/_instances-insert_profile_events-0-gw5/upstream_node/docker-compose.yml --file /ClickHouse/tests/integration/test_backward_compatibility/_instances-insert_profile_events-0-gw5/old_node/docker-compose.yml up -d --no-recreate') Stderr: 2130ef613416 Extracting [=====================> ] 110.3MB/258.8MB Command:[docker compose --env-file /ClickHouse/tests/integration/test_backward_compatibility/_instances-insert_profile_events-0-gw5/.env --project-name roottestbackwardcompatibilityinsertprofileevents-gw5 --file /ClickHouse/tests/integration/test_backward_compatibility/_instances-insert_profile_events-0-gw5/upstream_node/docker-compose.yml --file /ClickHouse/tests/integration/test_backward_compatibility/_instances-insert_profile_events-0-gw5/old_node/docker-compose.yml up -d --no-recreate] Stderr: 2130ef613416 Extracting [=======================> ] 120.3MB/258.8MB Stderr: 2130ef613416 Extracting [========================> ] 128.7MB/258.8MB Stderr: 2130ef613416 Extracting [==========================> ] 137.6MB/258.8MB Stderr: 2130ef613416 Extracting [============================> ] 148.2MB/258.8MB Stderr: 2130ef613416 Extracting [==============================> ] 157.1MB/258.8MB Stderr: 2130ef613416 Extracting [===============================> ] 165.4MB/258.8MB Stderr: 2130ef613416 Extracting [=================================> ] 172.1MB/258.8MB Stderr: 2130ef613416 Extracting [==================================> ] 181MB/258.8MB Stderr: 2130ef613416 Extracting [====================================> ] 191.1MB/258.8MB Stderr: 2130ef613416 Extracting [======================================> ] 200.5MB/258.8MB Stderr: 2130ef613416 Extracting [=========================================> ] 212.2MB/258.8MB Stderr: 2130ef613416 Extracting [==========================================> ] 221.7MB/258.8MB Stderr: 2130ef613416 Extracting [============================================> ] 229.5MB/258.8MB Stderr: 2130ef613416 Extracting [==============================================> ] 239.5MB/258.8MB Stderr: 2130ef613416 Extracting [================================================> ] 249.6MB/258.8MB Stderr: 2130ef613416 Extracting [=================================================> ] 257.4MB/258.8MB Stderr: 2130ef613416 Extracting [==================================================>] 258.8MB/258.8MB Stderr: 2130ef613416 Pull complete Stderr: 4f4fb700ef54 Extracting [==================================================>] 32B/32B Stderr: 4f4fb700ef54 Extracting [==================================================>] 32B/32B Stderr: 4f4fb700ef54 Pull complete Stderr: 6a3fe6bcecc5 Extracting [==================================================>] 22.46kB/22.46kB Stderr: 6a3fe6bcecc5 Extracting [==================================================>] 22.46kB/22.46kB Stderr: 6a3fe6bcecc5 Pull complete Stderr: 33bc578e5e14 Extracting [=> ] 32.77kB/863.5kB Stderr: 33bc578e5e14 Extracting [==================================================>] 863.5kB/863.5kB Stderr: 33bc578e5e14 Extracting [==================================================>] 863.5kB/863.5kB Stderr: 33bc578e5e14 Pull complete Stderr: 6d0df3aa878d Extracting [==================================================>] 116B/116B Stderr: 6d0df3aa878d Extracting [==================================================>] 116B/116B Stderr: 6d0df3aa878d Pull complete Stderr: 690edf6c102b Extracting [==================================================>] 364B/364B Stderr: 690edf6c102b Extracting [==================================================>] 364B/364B Stderr: 690edf6c102b Pull complete Stderr: 87e28718f844 Extracting [==================================================>] 2.936kB/2.936kB Stderr: 87e28718f844 Extracting [==================================================>] 2.936kB/2.936kB Stderr: 87e28718f844 Pull complete Stderr: backward Pulled ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_backward_compatibility/_instances-functions-0-gw9/.env --project-name roottestbackwardcompatibilityfunctions-gw9 --file /ClickHouse/tests/integration/test_backward_compatibility/_instances-functions-0-gw9/upstream/docker-compose.yml --file /ClickHouse/tests/integration/test_backward_compatibility/_instances-functions-0-gw9/backward/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_backward_compatibility/_instances-functions-0-gw9/.env --project-name roottestbackwardcompatibilityfunctions-gw9 --file /ClickHouse/tests/integration/test_backward_compatibility/_instances-functions-0-gw9/upstream/docker-compose.yml --file /ClickHouse/tests/integration/test_backward_compatibility/_instances-functions-0-gw9/backward/docker-compose.yml up -d --no-recreate] Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED get_instance_ip instance_name=zoo2 http://localhost:None "GET /v1.46/containers/roottestcompatibilitymergetreesettings-gw3-zoo2-1/json HTTP/1.1" 200 None get_kazoo_client: zoo2, ip:172.16.2.3, port:2181, use_ssl:False Connecting to 172.16.2.3(172.16.2.3):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED get_instance_ip instance_name=zoo3 http://localhost:None "GET /v1.46/containers/roottestcompatibilitymergetreesettings-gw3-zoo3-1/json HTTP/1.1" 200 None get_kazoo_client: zoo3, ip:172.16.2.4, port:2181, use_ssl:False Connecting to 172.16.2.4(172.16.2.4):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost Executing query SELECT status, error FROM system.backups WHERE id='5c6873f9-601e-4699-8235-c1bd3f21f1b4' on node1 Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED All instances of ZooKeeper started: ('zoo1', 'zoo2', 'zoo3') ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_compatibility_merge_tree_settings/_instances-0-gw3/.env --project-name roottestcompatibilitymergetreesettings-gw3 --file /ClickHouse/tests/integration/test_compatibility_merge_tree_settings/_instances-0-gw3/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_compatibility_merge_tree_settings/_instances-0-gw3/node2/docker-compose.yml --file /ClickHouse/tests/integration/test_compatibility_merge_tree_settings/_instances-0-gw3/node3/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_compatibility_merge_tree_settings/_instances-0-gw3/.env --project-name roottestcompatibilitymergetreesettings-gw3 --file /ClickHouse/tests/integration/test_compatibility_merge_tree_settings/_instances-0-gw3/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_compatibility_merge_tree_settings/_instances-0-gw3/node2/docker-compose.yml --file /ClickHouse/tests/integration/test_compatibility_merge_tree_settings/_instances-0-gw3/node3/docker-compose.yml up -d --no-recreate] Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster' mydb.tbl on node1 Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster' mydb.tbl2 on node1 Executing query SELECT * FROM mydb.tbl ORDER BY x on node1 Executing query SELECT * FROM mydb.tbl2 ORDER BY y on node2 Stderr: Container roottestcompatibilitymergetreesettings-gw3-zoo1-1 Running Stderr: Container roottestcompatibilitymergetreesettings-gw3-zoo2-1 Running Stderr: Container roottestcompatibilitymergetreesettings-gw3-zoo3-1 Running Stderr: Container roottestcompatibilitymergetreesettings-gw3-node1-1 Creating Stderr: Container roottestcompatibilitymergetreesettings-gw3-node2-1 Creating Stderr: Container roottestcompatibilitymergetreesettings-gw3-node3-1 Creating Stderr: Container roottestcompatibilitymergetreesettings-gw3-node3-1 Created Stderr: Container roottestcompatibilitymergetreesettings-gw3-node1-1 Created Stderr: Container roottestcompatibilitymergetreesettings-gw3-node2-1 Created Stderr: Container roottestcompatibilitymergetreesettings-gw3-node1-1 Starting Stderr: Container roottestcompatibilitymergetreesettings-gw3-node2-1 Starting Stderr: Container roottestcompatibilitymergetreesettings-gw3-node3-1 Starting Stderr: Container roottestcompatibilitymergetreesettings-gw3-node3-1 Started Stderr: Container roottestcompatibilitymergetreesettings-gw3-node1-1 Started Stderr: Container roottestcompatibilitymergetreesettings-gw3-node2-1 Started ClickHouse instance created get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottestcompatibilitymergetreesettings-gw3-node1-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottestcompatibilitymergetreesettings-gw3-node1-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node1, ip: 172.16.2.7... http://localhost:None "GET /v1.46/containers/roottestcompatibilitymergetreesettings-gw3-node1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/508f20571c736fee199dcb38f797a4c080749ac703b5f71cda9f7b48ab94e41b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/508f20571c736fee199dcb38f797a4c080749ac703b5f71cda9f7b48ab94e41b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/508f20571c736fee199dcb38f797a4c080749ac703b5f71cda9f7b48ab94e41b/json HTTP/1.1" 200 None Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw0] PASSED test_backup_restore_on_cluster/test.py::test_replicated_database_async http://localhost:None "GET /v1.46/containers/508f20571c736fee199dcb38f797a4c080749ac703b5f71cda9f7b48ab94e41b/json HTTP/1.1" 200 None Stderr: Container roottestalteronmixedtypecluster-gw7-node2-1 Stopping Stderr: Container roottestalteronmixedtypecluster-gw7-node3-1 Stopping Stderr: Container roottestalteronmixedtypecluster-gw7-node4-1 Stopping Stderr: Container roottestalteronmixedtypecluster-gw7-node1-1 Stopping Stderr: Container roottestalteronmixedtypecluster-gw7-node1-1 Stopped Stderr: Container roottestalteronmixedtypecluster-gw7-node4-1 Stopped Stderr: Container roottestalteronmixedtypecluster-gw7-node3-1 Stopped Stderr: Container roottestalteronmixedtypecluster-gw7-node2-1 Stopped Stderr: Container roottestalteronmixedtypecluster-gw7-zoo3-1 Stopping Stderr: Container roottestalteronmixedtypecluster-gw7-zoo1-1 Stopping Stderr: Container roottestalteronmixedtypecluster-gw7-zoo2-1 Stopping Stderr: Container roottestalteronmixedtypecluster-gw7-zoo2-1 Stopped Stderr: Container roottestalteronmixedtypecluster-gw7-zoo3-1 Stopped Stderr: Container roottestalteronmixedtypecluster-gw7-zoo1-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_alter_on_mixed_type_cluster/_instances-0-gw7/node1/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_alter_on_mixed_type_cluster/_instances-0-gw7/node1/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_alter_on_mixed_type_cluster/_instances-0-gw7/node2/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_alter_on_mixed_type_cluster/_instances-0-gw7/node2/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Command:[bash -c [ -f /ClickHouse/tests/integration/test_alter_on_mixed_type_cluster/_instances-0-gw7/node3/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_alter_on_mixed_type_cluster/_instances-0-gw7/node3/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_alter_on_mixed_type_cluster/_instances-0-gw7/node4/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_alter_on_mixed_type_cluster/_instances-0-gw7/node4/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] http://localhost:None "GET /v1.46/containers/508f20571c736fee199dcb38f797a4c080749ac703b5f71cda9f7b48ab94e41b/json HTTP/1.1" 200 None ClickHouse node1 started get_instance_ip instance_name=node2 Command:[docker compose --env-file /ClickHouse/tests/integration/test_alter_on_mixed_type_cluster/_instances-0-gw7/.env --project-name roottestalteronmixedtypecluster-gw7 --file /ClickHouse/tests/integration/test_alter_on_mixed_type_cluster/_instances-0-gw7/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_alter_on_mixed_type_cluster/_instances-0-gw7/node2/docker-compose.yml --file /ClickHouse/tests/integration/test_alter_on_mixed_type_cluster/_instances-0-gw7/node3/docker-compose.yml --file /ClickHouse/tests/integration/test_alter_on_mixed_type_cluster/_instances-0-gw7/node4/docker-compose.yml down --volumes] http://localhost:None "GET /v1.46/containers/roottestcompatibilitymergetreesettings-gw3-node2-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node2 http://localhost:None "GET /v1.46/containers/roottestcompatibilitymergetreesettings-gw3-node2-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node2, ip: 172.16.2.6... http://localhost:None "GET /v1.46/containers/roottestcompatibilitymergetreesettings-gw3-node2-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/02a02222a5f1f99a840700aa4f2bd050aa5d4232e9fe7a0ea8e214ce35557bfe/json HTTP/1.1" 200 None ClickHouse node2 started get_instance_ip instance_name=node3 http://localhost:None "GET /v1.46/containers/roottestcompatibilitymergetreesettings-gw3-node3-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node3 http://localhost:None "GET /v1.46/containers/roottestcompatibilitymergetreesettings-gw3-node3-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node3, ip: 172.16.2.5... http://localhost:None "GET /v1.46/containers/roottestcompatibilitymergetreesettings-gw3-node3-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/09cffa3d22fb1e87f3364b4ace1c6afa9e34666c33b4f591f563a54a33698303/json HTTP/1.1" 200 None ClickHouse node3 started Executing query CREATE TABLE tp (type Int32, eventcnt UInt64, PROJECTION p (select sum(eventcnt), type group by type)) engine = ReplacingMergeTree order by type; on node1 Stderr: Network roottestbackwardcompatibilityfunctions-gw9_default Creating Stderr: Network roottestbackwardcompatibilityinsertprofileevents-gw5_default Creating Stderr: Network roottestbackwardcompatibilityfunctions-gw9_default Created Stderr: Container roottestbackwardcompatibilityfunctions-gw9-backward-1 Creating Stderr: Container roottestbackwardcompatibilityfunctions-gw9-upstream-1 Creating Stderr: Container roottestbackwardcompatibilityfunctions-gw9-upstream-1 Created Stderr: Container roottestbackwardcompatibilityfunctions-gw9-backward-1 Created Stderr: Network roottestbackwardcompatibilityinsertprofileevents-gw5_default Created Stderr: Container roottestbackwardcompatibilityfunctions-gw9-upstream-1 Starting Stderr: Container roottestbackwardcompatibilityfunctions-gw9-backward-1 Starting Stderr: Container roottestbackwardcompatibilityfunctions-gw9-upstream-1 Started Stderr: Container roottestbackwardcompatibilityfunctions-gw9-backward-1 Started ClickHouse instance created get_instance_ip instance_name=upstream http://localhost:None "GET /v1.46/containers/roottestbackwardcompatibilityfunctions-gw9-upstream-1/json HTTP/1.1" 200 None get_instance_ip instance_name=upstream Stderr: Container roottestbackwardcompatibilityinsertprofileevents-gw5-upstream_node-1 Creating Stderr: Container roottestbackwardcompatibilityinsertprofileevents-gw5-old_node-1 Creating Stderr: Container roottestbackwardcompatibilityinsertprofileevents-gw5-upstream_node-1 Created Stderr: Container roottestbackwardcompatibilityinsertprofileevents-gw5-old_node-1 Created Stderr: Container roottestbackwardcompatibilityinsertprofileevents-gw5-upstream_node-1 Starting Stderr: Container roottestbackwardcompatibilityinsertprofileevents-gw5-old_node-1 Starting Stderr: Container roottestbackwardcompatibilityinsertprofileevents-gw5-old_node-1 Started Stderr: Container roottestbackwardcompatibilityinsertprofileevents-gw5-upstream_node-1 Started ClickHouse instance created get_instance_ip instance_name=upstream_node http://localhost:None "GET /v1.46/containers/roottestbackwardcompatibilityfunctions-gw9-upstream-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in upstream, ip: 172.16.5.3... http://localhost:None "GET /v1.46/containers/roottestbackwardcompatibilityinsertprofileevents-gw5-upstream_node-1/json HTTP/1.1" 200 None get_instance_ip instance_name=upstream_node http://localhost:None "GET /v1.46/containers/roottestbackwardcompatibilityfunctions-gw9-upstream-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestbackwardcompatibilityinsertprofileevents-gw5-upstream_node-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in upstream_node, ip: 172.16.4.3... http://localhost:None "GET /v1.46/containers/roottestbackwardcompatibilityinsertprofileevents-gw5-upstream_node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/0d7dfcb58d03e045ceab00f7bd079618787f9c3763e0c4aad2c8058c5eed9872/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/cb9f01252b35a39fa94a60e5868a9cf0b99cc33574324a54f339352ae3874eda/json HTTP/1.1" 200 None Executing query CREATE TABLE tp (type Int32, eventcnt UInt64, PROJECTION p (select sum(eventcnt), type group by type)) engine = ReplicatedReplacingMergeTree('/tables/tp', '0') order by type; on node1 http://localhost:None "GET /v1.46/containers/0d7dfcb58d03e045ceab00f7bd079618787f9c3763e0c4aad2c8058c5eed9872/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/cb9f01252b35a39fa94a60e5868a9cf0b99cc33574324a54f339352ae3874eda/json HTTP/1.1" 200 None Executing query CREATE TABLE tp (type Int32, eventcnt UInt64) engine = ReplacingMergeTree order by type; on node1 Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 http://localhost:None "GET /v1.46/containers/0d7dfcb58d03e045ceab00f7bd079618787f9c3763e0c4aad2c8058c5eed9872/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/cb9f01252b35a39fa94a60e5868a9cf0b99cc33574324a54f339352ae3874eda/json HTTP/1.1" 200 None Executing query ALTER TABLE tp ADD PROJECTION p (select sum(eventcnt), type group by type); on node1 Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Executing query drop table tp; on node1 http://localhost:None "GET /v1.46/containers/0d7dfcb58d03e045ceab00f7bd079618787f9c3763e0c4aad2c8058c5eed9872/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/cb9f01252b35a39fa94a60e5868a9cf0b99cc33574324a54f339352ae3874eda/json HTTP/1.1" 200 None Executing query CREATE TABLE tp (type Int32, eventcnt UInt64) engine = ReplicatedReplacingMergeTree('/tables/tp', '0') order by type; on node1 Stderr: Container roottestalteronmixedtypecluster-gw7-node4-1 Stopping Stderr: Container roottestalteronmixedtypecluster-gw7-node1-1 Stopping Stderr: Container roottestalteronmixedtypecluster-gw7-node2-1 Stopping Stderr: Container roottestalteronmixedtypecluster-gw7-node3-1 Stopping Stderr: Container roottestalteronmixedtypecluster-gw7-node1-1 Stopped Stderr: Container roottestalteronmixedtypecluster-gw7-node1-1 Removing Stderr: Container roottestalteronmixedtypecluster-gw7-node2-1 Stopped Stderr: Container roottestalteronmixedtypecluster-gw7-node2-1 Removing Stderr: Container roottestalteronmixedtypecluster-gw7-node3-1 Stopped Stderr: Container roottestalteronmixedtypecluster-gw7-node3-1 Removing Stderr: Container roottestalteronmixedtypecluster-gw7-node4-1 Stopped Stderr: Container roottestalteronmixedtypecluster-gw7-node4-1 Removing Stderr: Container roottestalteronmixedtypecluster-gw7-node1-1 Removed Stderr: Container roottestalteronmixedtypecluster-gw7-node4-1 Removed Stderr: Container roottestalteronmixedtypecluster-gw7-node3-1 Removed Stderr: Container roottestalteronmixedtypecluster-gw7-node2-1 Removed Stderr: Container roottestalteronmixedtypecluster-gw7-zoo1-1 Stopping Stderr: Container roottestalteronmixedtypecluster-gw7-zoo2-1 Stopping Stderr: Container roottestalteronmixedtypecluster-gw7-zoo3-1 Stopping Stderr: Container roottestalteronmixedtypecluster-gw7-zoo3-1 Stopped Stderr: Container roottestalteronmixedtypecluster-gw7-zoo3-1 Removing Stderr: Container roottestalteronmixedtypecluster-gw7-zoo1-1 Stopped Stderr: Container roottestalteronmixedtypecluster-gw7-zoo1-1 Removing Stderr: Container roottestalteronmixedtypecluster-gw7-zoo2-1 Stopped Stderr: Container roottestalteronmixedtypecluster-gw7-zoo2-1 Removing Stderr: Container roottestalteronmixedtypecluster-gw7-zoo2-1 Removed Stderr: Container roottestalteronmixedtypecluster-gw7-zoo1-1 Removed Stderr: Container roottestalteronmixedtypecluster-gw7-zoo3-1 Removed Stderr: Network roottestalteronmixedtypecluster-gw7_default Removing Stderr: Network roottestalteronmixedtypecluster-gw7_default Removed Cleanup called Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Docker networks for project roottestalteronmixedtypecluster-gw7 are NETWORK ID NAME DRIVER SCOPE Executing query ALTER TABLE tp ADD PROJECTION p (select sum(eventcnt), type group by type); on node1 http://localhost:None "GET /v1.46/containers/0d7dfcb58d03e045ceab00f7bd079618787f9c3763e0c4aad2c8058c5eed9872/json HTTP/1.1" 200 None Docker containers for project roottestalteronmixedtypecluster-gw7 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES ClickHouse upstream started get_instance_ip instance_name=backward http://localhost:None "GET /v1.46/containers/roottestbackwardcompatibilityfunctions-gw9-backward-1/json HTTP/1.1" 200 None get_instance_ip instance_name=backward http://localhost:None "GET /v1.46/containers/cb9f01252b35a39fa94a60e5868a9cf0b99cc33574324a54f339352ae3874eda/json HTTP/1.1" 200 None ClickHouse upstream_node started get_instance_ip instance_name=old_node http://localhost:None "GET /v1.46/containers/roottestbackwardcompatibilityfunctions-gw9-backward-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in backward, ip: 172.16.5.2... http://localhost:None "GET /v1.46/containers/roottestbackwardcompatibilityinsertprofileevents-gw5-old_node-1/json HTTP/1.1" 200 None get_instance_ip instance_name=old_node http://localhost:None "GET /v1.46/containers/roottestbackwardcompatibilityfunctions-gw9-backward-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestbackwardcompatibilityinsertprofileevents-gw5-old_node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/6dcc547c80fb69f516fbc05ea331d6c757178c8090135b6ff449302b85082cea/json HTTP/1.1" 200 None Waiting for ClickHouse start in old_node, ip: 172.16.4.2... ClickHouse backward started http://localhost:None "GET /v1.46/containers/roottestbackwardcompatibilityinsertprofileevents-gw5-old_node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/8df10164b16154e104ede18b3c3e58b843dde406695de0689ac466a252161104/json HTTP/1.1" 200 None Executing query SELECT if(NOT empty(alias_to), alias_to, name) FROM system.functions WHERE is_aggregate = 1 on backward ClickHouse old_node started Executing query INSERT INTO FUNCTION null('foo String') VALUES ('foo')('bar') on upstream_node Docker volumes for project roottestalteronmixedtypecluster-gw7 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestalteronmixedtypecluster-gw7-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestalteronmixedtypecluster-gw7 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Executing query drop table tp; on node1 Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Executing query INSERT INTO FUNCTION null('foo String') VALUES ('foo')('bar') on upstream_node Stdout:1 Volumes pruned: 1 test_buffer_profile/test.py::test_buffer_profile Running tests in /ClickHouse/tests/integration/test_buffer_profile/test.py Cluster start called. is_up=False Docker networks for project roottestbufferprofile-gw7 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestbufferprofile-gw7 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestbufferprofile-gw7 are DRIVER VOLUME NAME Cleanup called Executing query CREATE TABLE tp (type Int32, eventcnt UInt64, PROJECTION p (select sum(eventcnt), type group by type)) engine = ReplacingMergeTree order by type; on node2 Got 163 aggregate functions Checking exponentialTimeDecayedCount Executing query select hex(initializeAggregation('exponentialTimeDecayedCountState', 'foo')) on backward Docker networks for project roottestbufferprofile-gw7 are NETWORK ID NAME DRIVER SCOPE [gw5] PASSED test_backward_compatibility/test_insert_profile_events.py::test_new_client_compatible test_backward_compatibility/test_insert_profile_events.py::test_old_client_compatible Executing query INSERT INTO FUNCTION null('foo String') VALUES ('foo')('bar') on old_node Docker containers for project roottestbufferprofile-gw7 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestbufferprofile-gw7 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestbufferprofile-gw7-.*-1$' --format '{{.ID}}:{{.Names}}'] Executing query drop table tp; on node2 Skipping exponentialTimeDecayedCount Checking exponentialTimeDecayedMax Unstopped containers: {} Executing query select hex(initializeAggregation('exponentialTimeDecayedMaxState', 'foo')) on backward No running containers for project: roottestbufferprofile-gw7 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Executing query INSERT INTO FUNCTION null('foo String') VALUES ('foo')('bar') on old_node Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] test_backup_restore_on_cluster/test.py::test_replicated_database_compare_parts Executing query CREATE DATABASE mydb ON CLUSTER 'cluster' ENGINE=Replicated('/clickhouse/path/','{shard}','{replica}') on node1 Stdout:1 Volumes pruned: 1 Setup directory for instance: node_default Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files [] to /ClickHouse/tests/integration/test_buffer_profile/_instances-0-gw7/node_default/configs/config.d Setup database dir /ClickHouse/tests/integration/test_buffer_profile/_instances-0-gw7/node_default/database Setup logs dir /ClickHouse/tests/integration/test_buffer_profile/_instances-0-gw7/node_default/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Setup directory for instance: node_buffer_profile Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_buffer_profile/configs/buffer_profile.xml'] to /ClickHouse/tests/integration/test_buffer_profile/_instances-0-gw7/node_buffer_profile/configs/config.d Setup database dir /ClickHouse/tests/integration/test_buffer_profile/_instances-0-gw7/node_buffer_profile/database Setup logs dir /ClickHouse/tests/integration/test_buffer_profile/_instances-0-gw7/node_buffer_profile/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw'} stored in /ClickHouse/tests/integration/test_buffer_profile/_instances-0-gw7/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found http://localhost:None "GET /version HTTP/1.1" 200 826 Executing query CREATE TABLE tp (type Int32, eventcnt UInt64, PROJECTION p (select sum(eventcnt), type group by type)) engine = ReplicatedReplacingMergeTree('/tables/tp2', '0') order by type; on node2 Command:[docker compose --env-file /ClickHouse/tests/integration/test_buffer_profile/_instances-0-gw7/.env --project-name roottestbufferprofile-gw7 --file /ClickHouse/tests/integration/test_buffer_profile/_instances-0-gw7/node_default/docker-compose.yml --file /ClickHouse/tests/integration/test_buffer_profile/_instances-0-gw7/node_buffer_profile/docker-compose.yml pull] Skipping exponentialTimeDecayedMax Checking lagInFrame Executing query select hex(initializeAggregation('lagInFrameState', 'foo')) on backward Command:[docker compose --env-file /ClickHouse/tests/integration/test_backward_compatibility/_instances-insert_profile_events-0-gw5/.env --project-name roottestbackwardcompatibilityinsertprofileevents-gw5 --file /ClickHouse/tests/integration/test_backward_compatibility/_instances-insert_profile_events-0-gw5/upstream_node/docker-compose.yml --file /ClickHouse/tests/integration/test_backward_compatibility/_instances-insert_profile_events-0-gw5/old_node/docker-compose.yml stop --timeout 20] [gw5] PASSED test_backward_compatibility/test_insert_profile_events.py::test_old_client_compatible Executing query drop table tp; on node2 Skipping lagInFrame Checking dense_rank Executing query select hex(initializeAggregation('dense_rankState', 'foo')) on backward Executing query CREATE TABLE tp (type Int32, eventcnt UInt64) engine = ReplacingMergeTree order by type; on node2 Skipping dense_rank Checking rank Executing query select hex(initializeAggregation('rankState', 'foo')) on backward Executing query CREATE TABLE mydb.tbl(x UInt8, y String) ENGINE=ReplicatedMergeTree ORDER BY x on node1 Skipping rank Checking exponentialMovingAverage Executing query select hex(initializeAggregation('exponentialMovingAverageState', 'foo')) on backward Executing query ALTER TABLE tp ADD PROJECTION p (select sum(eventcnt), type group by type); on node2 Skipping exponentialMovingAverage Checking sparkbar Executing query select hex(initializeAggregation('sparkbarState', 'foo')) on backward Executing query drop table tp; on node2 Skipping sparkbar Checking singleValueOrNull Executing query select hex(initializeAggregation('singleValueOrNullState', 'foo')) on backward Executing query SYSTEM SYNC DATABASE REPLICA mydb on node2 Executing query CREATE TABLE tp (type Int32, eventcnt UInt64) engine = ReplicatedReplacingMergeTree('/tables/tp3', '0') order by type; on node2 Executing query select hex(initializeAggregation('singleValueOrNullState', 'foo')) on upstream Executing query SYSTEM STOP MERGES mydb.tbl on node1 Executing query SYSTEM STOP MERGES mydb.tbl on node2 Executing query ALTER TABLE tp ADD PROJECTION p (select sum(eventcnt), type group by type); on node2 OK singleValueOrNull Checking meanZTest Executing query select hex(initializeAggregation('meanZTestState', 'foo')) on backward Executing query SYSTEM STOP FETCHES mydb.tbl on node1 Executing query drop table tp; on node2 Skipping meanZTest Checking analysisOfVariance Executing query select hex(initializeAggregation('analysisOfVarianceState', 'foo')) on backward Executing query SYSTEM STOP FETCHES mydb.tbl on node2 [gw3] PASSED test_compatibility_merge_tree_settings/test.py::test_check_projections_compatibility test_compatibility_merge_tree_settings/test.py::test_config_overrides_compatibility Executing query CREATE TABLE tp (type Int32, eventcnt UInt64, PROJECTION p (select sum(eventcnt), type group by type)) engine = ReplacingMergeTree order by type; on node3 Skipping analysisOfVariance Checking studentTTest Executing query select hex(initializeAggregation('studentTTestState', 'foo')) on backward Executing query INSERT INTO mydb.tbl VALUES (1, 'a') on node1 Executing query CREATE TABLE tp (type Int32, eventcnt UInt64, PROJECTION p (select sum(eventcnt), type group by type)) engine = ReplicatedReplacingMergeTree('/tables/tp', '0') order by type; on node3 Skipping studentTTest Checking rankCorr Executing query select hex(initializeAggregation('rankCorrState', 'foo')) on backward Executing query INSERT INTO mydb.tbl VALUES (2, 'b') on node1 Command:[docker compose --env-file /ClickHouse/tests/integration/test_compatibility_merge_tree_settings/_instances-0-gw3/.env --project-name roottestcompatibilitymergetreesettings-gw3 --file /ClickHouse/tests/integration/test_compatibility_merge_tree_settings/_instances-0-gw3/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_compatibility_merge_tree_settings/_instances-0-gw3/node2/docker-compose.yml --file /ClickHouse/tests/integration/test_compatibility_merge_tree_settings/_instances-0-gw3/node3/docker-compose.yml stop --timeout 20] [gw3] PASSED test_compatibility_merge_tree_settings/test.py::test_config_overrides_compatibility Skipping rankCorr Checking aggThrow Executing query select hex(initializeAggregation('aggThrowState', 'foo')) on backward Executing query INSERT INTO mydb.tbl VALUES (3, 'x') on node2 Skipping aggThrow Checking categoricalInformationValue Executing query select hex(initializeAggregation('categoricalInformationValueState', 'foo')) on backward Executing query INSERT INTO mydb.tbl VALUES (4, 'y') on node2 Skipping categoricalInformationValue Checking groupArrayMovingAvg Executing query select hex(initializeAggregation('groupArrayMovingAvgState', 'foo')) on backward Executing query SELECT * FROM mydb.tbl ORDER BY x on node2 Skipping groupArrayMovingAvg Checking groupArrayMovingSum Executing query select hex(initializeAggregation('groupArrayMovingSumState', 'foo')) on backward Executing query BACKUP DATABASE mydb ON CLUSTER 'cluster' TO Disk('backups', '14') SETTINGS replica_num=2 on node1 Skipping groupArrayMovingSum Checking simpleLinearRegression Executing query select hex(initializeAggregation('simpleLinearRegressionState', 'foo')) on backward Skipping simpleLinearRegression Checking entropy Executing query select hex(initializeAggregation('entropyState', 'foo')) on backward Executing query select hex(initializeAggregation('entropyState', 'foo')) on upstream Executing query DROP DATABASE mydb ON CLUSTER 'cluster' SYNC on node1 OK entropy Checking histogram Executing query select hex(initializeAggregation('histogramState', 'foo')) on backward Skipping histogram Checking mannWhitneyUTest Executing query select hex(initializeAggregation('mannWhitneyUTestState', 'foo')) on backward Executing query RESTORE DATABASE mydb ON CLUSTER 'cluster' FROM Disk('backups', '14') SETTINGS replica_num_in_backup=2 on node1 Skipping mannWhitneyUTest Checking maxIntersections Executing query select hex(initializeAggregation('maxIntersectionsState', 'foo')) on backward Skipping maxIntersections Checking groupBitmapXor Executing query select hex(initializeAggregation('groupBitmapXorState', 'foo')) on backward Skipping groupBitmapXor Checking groupBitmapAnd Executing query select hex(initializeAggregation('groupBitmapAndState', 'foo')) on backward Skipping groupBitmapAnd Checking cramersVBiasCorrected Executing query select hex(initializeAggregation('cramersVBiasCorrectedState', 'foo')) on backward Skipping cramersVBiasCorrected Checking contingency Executing query select hex(initializeAggregation('contingencyState', 'foo')) on backward Skipping contingency Checking skewPop Executing query select hex(initializeAggregation('skewPopState', 'foo')) on backward Skipping skewPop Checking groupBitXor Executing query select hex(initializeAggregation('groupBitXorState', 'foo')) on backward Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster' mydb.tbl on node1 Skipping groupBitXor Checking groupBitOr Executing query select hex(initializeAggregation('groupBitOrState', 'foo')) on backward Skipping groupBitOr Checking exponentialTimeDecayedSum Executing query select hex(initializeAggregation('exponentialTimeDecayedSumState', 'foo')) on backward Executing query SELECT _part, * FROM mydb.tbl ORDER BY x on node1 Skipping exponentialTimeDecayedSum Checking topKWeighted Executing query select hex(initializeAggregation('topKWeightedState', 'foo')) on backward Skipping topKWeighted Checking stochasticLinearRegression Executing query select hex(initializeAggregation('stochasticLinearRegressionState', 'foo')) on backward Executing query SELECT _part, * FROM mydb.tbl ORDER BY x on node2 Skipping stochasticLinearRegression Checking corr Executing query select hex(initializeAggregation('corrState', 'foo')) on backward Executing query SELECT * FROM mydb.tbl ORDER BY x on node2 Skipping corr Checking uniqCombined64 Executing query select hex(initializeAggregation('uniqCombined64State', 'foo')) on backward Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw0] PASSED test_backup_restore_on_cluster/test.py::test_replicated_database_compare_parts Executing query select hex(initializeAggregation('uniqCombined64State', 'foo')) on upstream OK uniqCombined64 Checking intervalLengthSum Executing query select hex(initializeAggregation('intervalLengthSumState', 'foo')) on backward Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Skipping intervalLengthSum Checking nothing Executing query select hex(initializeAggregation('nothingState', 'foo')) on backward Executing query select hex(initializeAggregation('nothingState', 'foo')) on upstream Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 OK nothing Checking uniqCombined Executing query select hex(initializeAggregation('uniqCombinedState', 'foo')) on backward Executing query select hex(initializeAggregation('uniqCombinedState', 'foo')) on upstream Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 OK uniqCombined Checking sumMapFiltered Executing query select hex(initializeAggregation('sumMapFilteredState', 'foo')) on backward Skipping sumMapFiltered Checking minMappedArrays Executing query select hex(initializeAggregation('minMappedArraysState', 'foo')) on backward Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Skipping minMappedArrays Checking sumMappedArrays Executing query select hex(initializeAggregation('sumMappedArraysState', 'foo')) on backward Skipping sumMappedArrays Checking quantiles Executing query select hex(initializeAggregation('quantilesState', 'foo')) on backward test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[database] Executing query CREATE DATABASE mydb ON CLUSTER 'cluster' ENGINE=Replicated('/clickhouse/databases/{database}','{shard}','{replica}') on node1 Stderr: node2 Skipped - Image is already being pulled by node1 Stderr: node1 Pulling Stderr: node1 Pulled ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw6/.env --project-name roottestaggregationmemoryefficient-gw6 --file /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw6/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw6/node2/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw6/.env --project-name roottestaggregationmemoryefficient-gw6 --file /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw6/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw6/node2/docker-compose.yml up -d --no-recreate] Skipping quantiles Checking sum Executing query select hex(initializeAggregation('sumState', 'foo')) on backward Stderr: zoo2 Skipped - Image is already being pulled by zoo1 Stderr: zoo3 Skipped - Image is already being pulled by zoo1 Stderr: node2 Skipped - Image is already being pulled by zoo1 Stderr: node1 Skipped - Image is already being pulled by zoo1 Stderr: zoo1 Pulling Stderr: zoo1 Pulled Setup ZooKeeper Creating internal ZooKeeper dirs: ['/ClickHouse/tests/integration/test_async_connect_to_multiple_ips/_instances-0-gw8/keeper1/log', '/ClickHouse/tests/integration/test_async_connect_to_multiple_ips/_instances-0-gw8/keeper1/config', '/ClickHouse/tests/integration/test_async_connect_to_multiple_ips/_instances-0-gw8/keeper1/coordination', '/ClickHouse/tests/integration/test_async_connect_to_multiple_ips/_instances-0-gw8/keeper2/log', '/ClickHouse/tests/integration/test_async_connect_to_multiple_ips/_instances-0-gw8/keeper2/config', '/ClickHouse/tests/integration/test_async_connect_to_multiple_ips/_instances-0-gw8/keeper2/coordination', '/ClickHouse/tests/integration/test_async_connect_to_multiple_ips/_instances-0-gw8/keeper3/log', '/ClickHouse/tests/integration/test_async_connect_to_multiple_ips/_instances-0-gw8/keeper3/config', '/ClickHouse/tests/integration/test_async_connect_to_multiple_ips/_instances-0-gw8/keeper3/coordination'] Command:[docker compose --project-name roottestasyncconnecttomultipleips-gw8 --env-file /ClickHouse/tests/integration/test_async_connect_to_multiple_ips/_instances-0-gw8/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_net.yml --verbose up -d] Skipping sum Checking sumKahan Executing query select hex(initializeAggregation('sumKahanState', 'foo')) on backward Stderr: node_buffer_profile Skipped - Image is already being pulled by node_default Stderr: node_default Pulling Stderr: node_default Pulled ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_buffer_profile/_instances-0-gw7/.env --project-name roottestbufferprofile-gw7 --file /ClickHouse/tests/integration/test_buffer_profile/_instances-0-gw7/node_default/docker-compose.yml --file /ClickHouse/tests/integration/test_buffer_profile/_instances-0-gw7/node_buffer_profile/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_buffer_profile/_instances-0-gw7/.env --project-name roottestbufferprofile-gw7 --file /ClickHouse/tests/integration/test_buffer_profile/_instances-0-gw7/node_default/docker-compose.yml --file /ClickHouse/tests/integration/test_buffer_profile/_instances-0-gw7/node_buffer_profile/docker-compose.yml up -d --no-recreate] Executing query CREATE TABLE mydb.tbl(x Int64) ENGINE=ReplicatedMergeTree ORDER BY x on node1 Skipping sumKahan Checking corrMatrix Executing query select hex(initializeAggregation('corrMatrixState', 'foo')) on backward Stderr: Container roottestbackwardcompatibilityinsertprofileevents-gw5-old_node-1 Stopping Stderr: Container roottestbackwardcompatibilityinsertprofileevents-gw5-upstream_node-1 Stopping Stderr: Container roottestbackwardcompatibilityinsertprofileevents-gw5-old_node-1 Stopped Stderr: Container roottestbackwardcompatibilityinsertprofileevents-gw5-upstream_node-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_backward_compatibility/_instances-insert_profile_events-0-gw5/upstream_node/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_backward_compatibility/_instances-insert_profile_events-0-gw5/upstream_node/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_backward_compatibility/_instances-insert_profile_events-0-gw5/old_node/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_backward_compatibility/_instances-insert_profile_events-0-gw5/old_node/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_backward_compatibility/_instances-insert_profile_events-0-gw5/.env --project-name roottestbackwardcompatibilityinsertprofileevents-gw5 --file /ClickHouse/tests/integration/test_backward_compatibility/_instances-insert_profile_events-0-gw5/upstream_node/docker-compose.yml --file /ClickHouse/tests/integration/test_backward_compatibility/_instances-insert_profile_events-0-gw5/old_node/docker-compose.yml down --volumes] Skipping corrMatrix Checking quantileExactWeighted Executing query select hex(initializeAggregation('quantileExactWeightedState', 'foo')) on backward Skipping quantileExactWeighted Checking sumMapFilteredWithOverflow Executing query select hex(initializeAggregation('sumMapFilteredWithOverflowState', 'foo')) on backward Executing query INSERT INTO mydb.tbl VALUES (-3) on node1 Skipping sumMapFilteredWithOverflow Checking corrStable Executing query select hex(initializeAggregation('corrStableState', 'foo')) on backward Executing query INSERT INTO mydb.tbl VALUES (1) on node1 Skipping corrStable Checking covarPopMatrix Executing query select hex(initializeAggregation('covarPopMatrixState', 'foo')) on backward Executing query INSERT INTO mydb.tbl VALUES (10) on node1 Skipping covarPopMatrix Checking groupBitmapOr Executing query select hex(initializeAggregation('groupBitmapOrState', 'foo')) on backward Executing query BACKUP DATABASE mydb ON CLUSTER 'cluster' TO Disk('backups', '15') on node1 Skipping groupBitmapOr Checking covarSampMatrix Executing query select hex(initializeAggregation('covarSampMatrixState', 'foo')) on backward Stderr: Network roottestaggregationmemoryefficient-gw6_default Creating Stderr: Network roottestaggregationmemoryefficient-gw6_default Created Stderr: Container roottestaggregationmemoryefficient-gw6-node2-1 Creating Stderr: Container roottestaggregationmemoryefficient-gw6-node1-1 Creating Stderr: Container roottestaggregationmemoryefficient-gw6-node1-1 Created Stderr: Container roottestaggregationmemoryefficient-gw6-node2-1 Created Stderr: Container roottestaggregationmemoryefficient-gw6-node2-1 Starting Stderr: Container roottestaggregationmemoryefficient-gw6-node1-1 Starting Stderr: Container roottestaggregationmemoryefficient-gw6-node2-1 Started Stderr: Container roottestaggregationmemoryefficient-gw6-node1-1 Started ClickHouse instance created get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottestaggregationmemoryefficient-gw6-node1-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottestaggregationmemoryefficient-gw6-node1-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node1, ip: 172.16.6.3... http://localhost:None "GET /v1.46/containers/roottestaggregationmemoryefficient-gw6-node1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/1ee8408afc4be505d941c20a382166cb58b0e5021568650c5396b1caa457c589/json HTTP/1.1" 200 None Executing query RESTORE DATABASE mydb AS mydb2 ON CLUSTER 'cluster' FROM Disk('backups', '15') on node1 Skipping covarSampMatrix Checking covarPop Executing query select hex(initializeAggregation('covarPopState', 'foo')) on backward http://localhost:None "GET /v1.46/containers/1ee8408afc4be505d941c20a382166cb58b0e5021568650c5396b1caa457c589/json HTTP/1.1" 200 None Skipping covarPop Checking row_number Executing query select hex(initializeAggregation('row_numberState', 'foo')) on backward http://localhost:None "GET /v1.46/containers/1ee8408afc4be505d941c20a382166cb58b0e5021568650c5396b1caa457c589/json HTTP/1.1" 200 None Stderr: Container roottestbackwardcompatibilityinsertprofileevents-gw5-old_node-1 Stopping Stderr: Container roottestbackwardcompatibilityinsertprofileevents-gw5-upstream_node-1 Stopping Stderr: Container roottestbackwardcompatibilityinsertprofileevents-gw5-old_node-1 Stopped Stderr: Container roottestbackwardcompatibilityinsertprofileevents-gw5-old_node-1 Removing Stderr: Container roottestbackwardcompatibilityinsertprofileevents-gw5-upstream_node-1 Stopped Stderr: Container roottestbackwardcompatibilityinsertprofileevents-gw5-upstream_node-1 Removing Stderr: Container roottestbackwardcompatibilityinsertprofileevents-gw5-upstream_node-1 Removed Stderr: Container roottestbackwardcompatibilityinsertprofileevents-gw5-old_node-1 Removed Stderr: Network roottestbackwardcompatibilityinsertprofileevents-gw5_default Removing Stderr: Network roottestbackwardcompatibilityinsertprofileevents-gw5_default Removed Cleanup called http://localhost:None "GET /v1.46/containers/1ee8408afc4be505d941c20a382166cb58b0e5021568650c5396b1caa457c589/json HTTP/1.1" 200 None Docker networks for project roottestbackwardcompatibilityinsertprofileevents-gw5 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestbackwardcompatibilityinsertprofileevents-gw5 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestbackwardcompatibilityinsertprofileevents-gw5 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestbackwardcompatibilityinsertprofileevents-gw5-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestbackwardcompatibilityinsertprofileevents-gw5 Trying to prune unused networks... Executing query INSERT INTO mydb.tbl VALUES (2) on node1 Trying to prune unused images... Command:[docker image prune -f] http://localhost:None "GET /v1.46/containers/1ee8408afc4be505d941c20a382166cb58b0e5021568650c5396b1caa457c589/json HTTP/1.1" 200 None ClickHouse node1 started get_instance_ip instance_name=node2 http://localhost:None "GET /v1.46/containers/roottestaggregationmemoryefficient-gw6-node2-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node2 http://localhost:None "GET /v1.46/containers/roottestaggregationmemoryefficient-gw6-node2-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node2, ip: 172.16.6.2... http://localhost:None "GET /v1.46/containers/roottestaggregationmemoryefficient-gw6-node2-1/json HTTP/1.1" 200 None Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] http://localhost:None "GET /v1.46/containers/d9e6e9dc228dab09acf6874612a8ab72827571b2717827f084659d3e81828835/json HTTP/1.1" 200 None ClickHouse node2 started Executing query create table da_memory_efficient_shard(A Int64, B Int64) Engine=MergeTree order by A partition by B % 2; on node1 Stdout:1 Volumes pruned: 1 test_backward_compatibility/test_convert_ordinary.py::test_convert_ordinary_to_atomic Running tests in /ClickHouse/tests/integration/test_backward_compatibility/test_convert_ordinary.py Cluster start called. is_up=False Docker networks for project roottestbackwardcompatibilityconvertordinary-gw5 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestbackwardcompatibilityconvertordinary-gw5 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestbackwardcompatibilityconvertordinary-gw5 are DRIVER VOLUME NAME Cleanup called Executing query SYSTEM SYNC DATABASE REPLICA ON CLUSTER 'cluster' mydb2 on node1 Docker networks for project roottestbackwardcompatibilityconvertordinary-gw5 are NETWORK ID NAME DRIVER SCOPE Skipping row_number Checking kurtPop Executing query select hex(initializeAggregation('kurtPopState', 'foo')) on backward Executing query create table da_memory_efficient_shard(A Int64, B Int64) Engine=MergeTree order by A partition by B % 2; on node2 Docker containers for project roottestbackwardcompatibilityconvertordinary-gw5 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Stderr:time="2025-04-02T03:34:31Z" level=trace msg="Docker Desktop integration not enabled" Stderr: Network roottestasyncconnecttomultipleips-gw8_default Creating Stderr: Network roottestasyncconnecttomultipleips-gw8_default Created Stderr: Container roottestasyncconnecttomultipleips-gw8-zoo2-1 Creating Stderr: Container roottestasyncconnecttomultipleips-gw8-zoo1-1 Creating Stderr: Container roottestasyncconnecttomultipleips-gw8-zoo3-1 Creating Stderr: Container roottestasyncconnecttomultipleips-gw8-zoo3-1 Created Stderr: Container roottestasyncconnecttomultipleips-gw8-zoo2-1 Created Stderr: Container roottestasyncconnecttomultipleips-gw8-zoo1-1 Created Stderr: Container roottestasyncconnecttomultipleips-gw8-zoo1-1 Starting Stderr: Container roottestasyncconnecttomultipleips-gw8-zoo2-1 Starting Stderr: Container roottestasyncconnecttomultipleips-gw8-zoo3-1 Starting Stderr: Container roottestasyncconnecttomultipleips-gw8-zoo3-1 Started Stderr: Container roottestasyncconnecttomultipleips-gw8-zoo2-1 Started Stderr: Container roottestasyncconnecttomultipleips-gw8-zoo1-1 Started Stderr:time="2025-04-02T03:34:32Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T03:34:32Z" level=debug msg="otel error" error="" Wait ZooKeeper to start get_instance_ip instance_name=zoo1 http://localhost:None "GET /v1.46/containers/roottestasyncconnecttomultipleips-gw8-zoo1-1/json HTTP/1.1" 200 None get_kazoo_client: zoo1, ip:10.0.0.4, port:2181, use_ssl:False Connecting to 10.0.0.4(10.0.0.4):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Docker volumes for project roottestbackwardcompatibilityconvertordinary-gw5 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestbackwardcompatibilityconvertordinary-gw5-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestbackwardcompatibilityconvertordinary-gw5 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:1 Volumes pruned: 1 Setup directory for instance: node Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files [] to /ClickHouse/tests/integration/test_backward_compatibility/_instances-convert_ordinary-0-gw5/node/configs/config.d Setup database dir /ClickHouse/tests/integration/test_backward_compatibility/_instances-convert_ordinary-0-gw5/node/database Setup logs dir /ClickHouse/tests/integration/test_backward_compatibility/_instances-convert_ordinary-0-gw5/node/logs Entrypoint cmd: bash -c "trap 'pkill tail' INT TERM; clickhouse server --config-file=/etc/clickhouse-server/config.xml --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log --daemon -- ; coproc tail -f /dev/null; wait $$!" Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'keeper_binary': '/clickhouse', 'keeper_cmd_prefix': 'clickhouse keeper', 'image': 'altinityinfra/integration-test:8b2301119731', 'user': '0', 'keeper_fs': 'bind', 'keeper_logs_dir1': '/ClickHouse/tests/integration/test_backward_compatibility/_instances-convert_ordinary-0-gw5/keeper1/log', 'keeper_config_dir1': '/ClickHouse/tests/integration/test_backward_compatibility/_instances-convert_ordinary-0-gw5/keeper1/config', 'keeper_db_dir1': '/ClickHouse/tests/integration/test_backward_compatibility/_instances-convert_ordinary-0-gw5/keeper1/coordination', 'keeper_logs_dir2': '/ClickHouse/tests/integration/test_backward_compatibility/_instances-convert_ordinary-0-gw5/keeper2/log', 'keeper_config_dir2': '/ClickHouse/tests/integration/test_backward_compatibility/_instances-convert_ordinary-0-gw5/keeper2/config', 'keeper_db_dir2': '/ClickHouse/tests/integration/test_backward_compatibility/_instances-convert_ordinary-0-gw5/keeper2/coordination', 'keeper_logs_dir3': '/ClickHouse/tests/integration/test_backward_compatibility/_instances-convert_ordinary-0-gw5/keeper3/log', 'keeper_config_dir3': '/ClickHouse/tests/integration/test_backward_compatibility/_instances-convert_ordinary-0-gw5/keeper3/config', 'keeper_db_dir3': '/ClickHouse/tests/integration/test_backward_compatibility/_instances-convert_ordinary-0-gw5/keeper3/coordination'} stored in /ClickHouse/tests/integration/test_backward_compatibility/_instances-convert_ordinary-0-gw5/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Connecting to 10.0.0.4(10.0.0.4):2181, use_ssl: False Connection dropped: socket connection error: Connection refused http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_backward_compatibility/_instances-convert_ordinary-0-gw5/.env --project-name roottestbackwardcompatibilityconvertordinary-gw5 --file /ClickHouse/tests/integration/test_backward_compatibility/_instances-convert_ordinary-0-gw5/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml pull] Stderr: Network roottestbufferprofile-gw7_default Creating Stderr: Network roottestbufferprofile-gw7_default Created Stderr: Container roottestbufferprofile-gw7-node_buffer_profile-1 Creating Stderr: Container roottestbufferprofile-gw7-node_default-1 Creating Stderr: Container roottestbufferprofile-gw7-node_default-1 Created Stderr: Container roottestbufferprofile-gw7-node_buffer_profile-1 Created Stderr: Container roottestbufferprofile-gw7-node_default-1 Starting Stderr: Container roottestbufferprofile-gw7-node_buffer_profile-1 Starting Stderr: Container roottestbufferprofile-gw7-node_default-1 Started Stderr: Container roottestbufferprofile-gw7-node_buffer_profile-1 Started ClickHouse instance created get_instance_ip instance_name=node_default http://localhost:None "GET /v1.46/containers/roottestbufferprofile-gw7-node_default-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node_default http://localhost:None "GET /v1.46/containers/roottestbufferprofile-gw7-node_default-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node_default, ip: 172.16.7.2... http://localhost:None "GET /v1.46/containers/roottestbufferprofile-gw7-node_default-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/615154bc0b17572e85804af64db02ef54ad6a7f1b06df5c98dd151cded82e8c3/json HTTP/1.1" 200 None Skipping kurtPop Checking kurtSamp Executing query select hex(initializeAggregation('kurtSampState', 'foo')) on backward Executing query insert into da_memory_efficient_shard select number, number from numbers(100000); on node1 Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster' mydb2.tbl on node1 Skipping kurtSamp Checking cramersV Executing query select hex(initializeAggregation('cramersVState', 'foo')) on backward http://localhost:None "GET /v1.46/containers/615154bc0b17572e85804af64db02ef54ad6a7f1b06df5c98dd151cded82e8c3/json HTTP/1.1" 200 None Connecting to 10.0.0.4(10.0.0.4):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Executing query insert into da_memory_efficient_shard select number + 100000, number from numbers(100000); on node2 Skipping cramersV Checking skewSamp Executing query select hex(initializeAggregation('skewSampState', 'foo')) on backward Executing query SELECT * FROM mydb.tbl ORDER BY x on node1 http://localhost:None "GET /v1.46/containers/615154bc0b17572e85804af64db02ef54ad6a7f1b06df5c98dd151cded82e8c3/json HTTP/1.1" 200 None Skipping skewSamp Checking nonNegativeDerivative Executing query select hex(initializeAggregation('nonNegativeDerivativeState', 'foo')) on backward Executing query SELECT * FROM mydb2.tbl ORDER BY x on node1 Executing query set distributed_aggregation_memory_efficient = 1, group_by_two_level_threshold = 1, group_by_two_level_threshold_bytes=1 on node1 Skipping nonNegativeDerivative Checking uniqExact Executing query select hex(initializeAggregation('uniqExactState', 'foo')) on backward Executing query SELECT * FROM mydb2.tbl ORDER BY x on node2 http://localhost:None "GET /v1.46/containers/615154bc0b17572e85804af64db02ef54ad6a7f1b06df5c98dd151cded82e8c3/json HTTP/1.1" 200 None Executing query select sum(a) from (SELECT B, uniqExact(A) a FROM remote('node{1,2}', default.da_memory_efficient_shard) GROUP BY B) on node1 Executing query select hex(initializeAggregation('uniqExactState', 'foo')) on upstream Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw0] PASSED test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[database] OK uniqExact Checking sumMapWithOverflow Executing query select hex(initializeAggregation('sumMapWithOverflowState', 'foo')) on backward http://localhost:None "GET /v1.46/containers/615154bc0b17572e85804af64db02ef54ad6a7f1b06df5c98dd151cded82e8c3/json HTTP/1.1" 200 None ClickHouse node_default started get_instance_ip instance_name=node_buffer_profile http://localhost:None "GET /v1.46/containers/roottestbufferprofile-gw7-node_buffer_profile-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node_buffer_profile http://localhost:None "GET /v1.46/containers/roottestbufferprofile-gw7-node_buffer_profile-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node_buffer_profile, ip: 172.16.7.3... http://localhost:None "GET /v1.46/containers/roottestbufferprofile-gw7-node_buffer_profile-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/a7834b83f4de3af98c8dee0bcd11b1acabd4514d3b7935c064992657921e0161/json HTTP/1.1" 200 None ClickHouse node_buffer_profile started Executing query CREATE TABLE data (key Int) Engine=MergeTree() ORDER BY key PARTITION BY key % 2; CREATE TABLE buffer AS data Engine=Buffer(currentDatabase(), data, /* settings for manual flush only */ 1, /* num_layers */ 10e6, /* min_time, placeholder */ 10e6, /* max_time, placeholder */ 0, /* min_rows */ 10e6, /* max_rows */ 0, /* min_bytes */ 80e6 /* max_bytes */ ); INSERT INTO buffer SELECT * FROM numbers(100); on node_buffer_profile Connecting to 10.0.0.4(10.0.0.4):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Skipping sumMapWithOverflow Checking stddevSamp Executing query select hex(initializeAggregation('stddevSampState', 'foo')) on backward Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Executing query set distributed_aggregation_memory_efficient = 0 on node1 Skipping stddevSamp Checking varPop Executing query select hex(initializeAggregation('varPopState', 'foo')) on backward Executing query OPTIMIZE TABLE buffer on node_buffer_profile Executing query select sum(a) from (SELECT B, uniqExact(A) a FROM remote('node{1,2}', default.da_memory_efficient_shard) GROUP BY B) on node1 Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Skipping varPop Checking quantileTimingWeighted Executing query select hex(initializeAggregation('quantileTimingWeightedState', 'foo')) on backward [gw7] PASSED test_buffer_profile/test.py::test_buffer_profile test_buffer_profile/test.py::test_default_profile Executing query CREATE TABLE data (key Int) Engine=MergeTree() ORDER BY key PARTITION BY key % 2; CREATE TABLE buffer AS data Engine=Buffer(currentDatabase(), data, /* settings for manual flush only */ 1, /* num_layers */ 10e6, /* min_time, placeholder */ 10e6, /* max_time, placeholder */ 0, /* min_rows */ 10e6, /* max_rows */ 0, /* min_bytes */ 80e6 /* max_bytes */ ); INSERT INTO buffer SELECT * FROM numbers(100); on node_default Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Skipping quantileTimingWeighted Checking covarPopStable Executing query select hex(initializeAggregation('covarPopStableState', 'foo')) on backward Executing query set distributed_aggregation_memory_efficient = 1, group_by_two_level_threshold = 1, group_by_two_level_threshold_bytes=1 on node1 Executing query OPTIMIZE TABLE buffer on node_default Skipping covarPopStable Checking stddevSampStable Executing query select hex(initializeAggregation('stddevSampStableState', 'foo')) on backward Connecting to 10.0.0.4(10.0.0.4):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Skipping stddevSampStable Checking varSamp Executing query select hex(initializeAggregation('varSampState', 'foo')) on backward Executing query SELECT fullHostName() AS h, uniqExact(A) AS a FROM remote('node{1,2}', default.da_memory_efficient_shard) GROUP BY h ORDER BY h; on node1 Command:[docker compose --env-file /ClickHouse/tests/integration/test_buffer_profile/_instances-0-gw7/.env --project-name roottestbufferprofile-gw7 --file /ClickHouse/tests/integration/test_buffer_profile/_instances-0-gw7/node_default/docker-compose.yml --file /ClickHouse/tests/integration/test_buffer_profile/_instances-0-gw7/node_buffer_profile/docker-compose.yml stop --timeout 20] [gw7] PASSED test_buffer_profile/test.py::test_default_profile Skipping varSamp Checking topK Executing query select hex(initializeAggregation('topKState', 'foo')) on backward Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Executing query select hex(initializeAggregation('topKState', 'foo')) on upstream Executing query set distributed_aggregation_memory_efficient = 0 on node1 OK topK Checking last_value Executing query select hex(initializeAggregation('last_valueState', 'foo')) on backward Executing query SELECT fullHostName() AS h, uniqExact(A) AS a FROM remote('node{1,2}', default.da_memory_efficient_shard) GROUP BY h ORDER BY h; on node1 test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[uuid] Executing query CREATE DATABASE mydb ON CLUSTER 'cluster' ENGINE=Replicated('/clickhouse/databases/{uuid}','{shard}','{replica}') on node1 Executing query select hex(initializeAggregation('last_valueState', 'foo')) on upstream OK last_value Checking any Executing query select hex(initializeAggregation('anyState', 'foo')) on backward Command:[docker compose --env-file /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw6/.env --project-name roottestaggregationmemoryefficient-gw6 --file /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw6/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw6/node2/docker-compose.yml stop --timeout 20] [gw6] PASSED test_aggregation_memory_efficient/test.py::test_remote Executing query CREATE TABLE mydb.tbl(x Int64) ENGINE=ReplicatedMergeTree ORDER BY x on node1 Executing query select hex(initializeAggregation('anyState', 'foo')) on upstream OK any Checking anyLast Executing query select hex(initializeAggregation('anyLastState', 'foo')) on backward Executing query select hex(initializeAggregation('anyLastState', 'foo')) on upstream Executing query INSERT INTO mydb.tbl VALUES (-3) on node1 OK anyLast Checking deltaSumTimestamp Executing query select hex(initializeAggregation('deltaSumTimestampState', 'foo')) on backward Executing query INSERT INTO mydb.tbl VALUES (1) on node1 Skipping deltaSumTimestamp Checking varSampStable Executing query select hex(initializeAggregation('varSampStableState', 'foo')) on backward Executing query INSERT INTO mydb.tbl VALUES (10) on node1 Skipping varSampStable Checking nth_value Executing query select hex(initializeAggregation('nth_valueState', 'foo')) on backward Executing query BACKUP DATABASE mydb ON CLUSTER 'cluster' TO Disk('backups', '16') on node1 Skipping nth_value Checking argMax Executing query select hex(initializeAggregation('argMaxState', 'foo')) on backward Skipping argMax Checking quantilesExact Executing query select hex(initializeAggregation('quantilesExactState', 'foo')) on backward Executing query RESTORE DATABASE mydb AS mydb2 ON CLUSTER 'cluster' FROM Disk('backups', '16') on node1 Skipping quantilesExact Checking uniqHLL12 Executing query select hex(initializeAggregation('uniqHLL12State', 'foo')) on backward Executing query select hex(initializeAggregation('uniqHLL12State', 'foo')) on upstream OK uniqHLL12 Checking quantileBFloat16 Executing query select hex(initializeAggregation('quantileBFloat16State', 'foo')) on backward Skipping quantileBFloat16 Checking uniq Executing query select hex(initializeAggregation('uniqState', 'foo')) on backward Executing query select hex(initializeAggregation('uniqState', 'foo')) on upstream OK uniq Checking min Executing query select hex(initializeAggregation('minState', 'foo')) on backward Executing query INSERT INTO mydb.tbl VALUES (2) on node1 Executing query select hex(initializeAggregation('minState', 'foo')) on upstream Executing query SYSTEM SYNC DATABASE REPLICA ON CLUSTER 'cluster' mydb2 on node1 OK min Checking sequenceNextNode Executing query select hex(initializeAggregation('sequenceNextNodeState', 'foo')) on backward Skipping sequenceNextNode Checking quantilesTimingWeighted Executing query select hex(initializeAggregation('quantilesTimingWeightedState', 'foo')) on backward Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster' mydb2.tbl on node1 Skipping quantilesTimingWeighted Checking ntile Executing query select hex(initializeAggregation('ntileState', 'foo')) on backward Skipping ntile Checking boundingRatio Executing query select hex(initializeAggregation('boundingRatioState', 'foo')) on backward Executing query SELECT * FROM mydb.tbl ORDER BY x on node1 Skipping boundingRatio Checking deltaSum Executing query select hex(initializeAggregation('deltaSumState', 'foo')) on backward Executing query SELECT * FROM mydb2.tbl ORDER BY x on node1 Skipping deltaSum Checking uniqUpTo Executing query select hex(initializeAggregation('uniqUpToState', 'foo')) on backward Executing query SELECT * FROM mydb2.tbl ORDER BY x on node2 Connecting to 10.0.0.4(10.0.0.4):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost Executing query select hex(initializeAggregation('uniqUpToState', 'foo')) on upstream Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw0] PASSED test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[uuid] Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED get_instance_ip instance_name=zoo2 http://localhost:None "GET /v1.46/containers/roottestasyncconnecttomultipleips-gw8-zoo2-1/json HTTP/1.1" 200 None get_kazoo_client: zoo2, ip:10.0.0.2, port:2181, use_ssl:False Connecting to 10.0.0.2(10.0.0.2):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost OK uniqUpTo Checking windowFunnel Executing query select hex(initializeAggregation('windowFunnelState', 'foo')) on backward Skipping windowFunnel Checking retention Executing query select hex(initializeAggregation('retentionState', 'foo')) on backward Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED get_instance_ip instance_name=zoo3 http://localhost:None "GET /v1.46/containers/roottestasyncconnecttomultipleips-gw8-zoo3-1/json HTTP/1.1" 200 None get_kazoo_client: zoo3, ip:10.0.0.3, port:2181, use_ssl:False Connecting to 10.0.0.3(10.0.0.3):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Skipping retention Checking sequenceMatch Executing query select hex(initializeAggregation('sequenceMatchState', 'foo')) on backward Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED All instances of ZooKeeper started: ('zoo1', 'zoo2', 'zoo3') ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_async_connect_to_multiple_ips/_instances-0-gw8/.env --project-name roottestasyncconnecttomultipleips-gw8 --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_net.yml --file /ClickHouse/tests/integration/test_async_connect_to_multiple_ips/_instances-0-gw8/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_async_connect_to_multiple_ips/_instances-0-gw8/node2/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_async_connect_to_multiple_ips/_instances-0-gw8/.env --project-name roottestasyncconnecttomultipleips-gw8 --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_net.yml --file /ClickHouse/tests/integration/test_async_connect_to_multiple_ips/_instances-0-gw8/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_async_connect_to_multiple_ips/_instances-0-gw8/node2/docker-compose.yml up -d --no-recreate] Skipping sequenceMatch Checking uniqTheta Executing query select hex(initializeAggregation('uniqThetaState', 'foo')) on backward Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Executing query select hex(initializeAggregation('uniqThetaState', 'foo')) on upstream Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 OK uniqTheta Checking quantilesExactWeighted Executing query select hex(initializeAggregation('quantilesExactWeightedState', 'foo')) on backward Skipping quantilesExactWeighted Checking max Executing query select hex(initializeAggregation('maxState', 'foo')) on backward Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Stderr: Container roottestasyncconnecttomultipleips-gw8-zoo1-1 Running Stderr: Container roottestasyncconnecttomultipleips-gw8-zoo2-1 Running Stderr: Container roottestasyncconnecttomultipleips-gw8-zoo3-1 Running Stderr: Container roottestasyncconnecttomultipleips-gw8-node1-1 Creating Stderr: Container roottestasyncconnecttomultipleips-gw8-node2-1 Creating Stderr: Container roottestasyncconnecttomultipleips-gw8-node1-1 Created Stderr: Container roottestasyncconnecttomultipleips-gw8-node2-1 Created Stderr: Container roottestasyncconnecttomultipleips-gw8-node1-1 Starting Stderr: Container roottestasyncconnecttomultipleips-gw8-node2-1 Starting Stderr: Container roottestasyncconnecttomultipleips-gw8-node2-1 Started Stderr: Container roottestasyncconnecttomultipleips-gw8-node1-1 Started ClickHouse instance created get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottestasyncconnecttomultipleips-gw8-node1-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottestasyncconnecttomultipleips-gw8-node1-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node1, ip: 10.5.95.11... http://localhost:None "GET /v1.46/containers/roottestasyncconnecttomultipleips-gw8-node1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/667c290c56ac5366bd1cc9b7162904c9500d7e17104828be4187f0cac5cc7226/json HTTP/1.1" 200 None Executing query select hex(initializeAggregation('maxState', 'foo')) on upstream OK max Checking quantilesBFloat16Weighted Executing query select hex(initializeAggregation('quantilesBFloat16WeightedState', 'foo')) on backward test_backup_restore_on_cluster/test.py::test_replicated_table Executing query CREATE TABLE tbl ON CLUSTER 'cluster' (x UInt8, y String) ENGINE=ReplicatedMergeTree('/clickhouse/tables/tbl/', '{replica}')ORDER BY x on node1 http://localhost:None "GET /v1.46/containers/667c290c56ac5366bd1cc9b7162904c9500d7e17104828be4187f0cac5cc7226/json HTTP/1.1" 200 None Skipping quantilesBFloat16Weighted Checking quantileBFloat16Weighted Executing query select hex(initializeAggregation('quantileBFloat16WeightedState', 'foo')) on backward http://localhost:None "GET /v1.46/containers/667c290c56ac5366bd1cc9b7162904c9500d7e17104828be4187f0cac5cc7226/json HTTP/1.1" 200 None Executing query INSERT INTO tbl VALUES (1, 'Don''t') on node1 http://localhost:None "GET /v1.46/containers/667c290c56ac5366bd1cc9b7162904c9500d7e17104828be4187f0cac5cc7226/json HTTP/1.1" 200 None Skipping quantileBFloat16Weighted Checking quantilesTDigestWeighted Executing query select hex(initializeAggregation('quantilesTDigestWeightedState', 'foo')) on backward Executing query INSERT INTO tbl VALUES (2, 'count') on node2 http://localhost:None "GET /v1.46/containers/667c290c56ac5366bd1cc9b7162904c9500d7e17104828be4187f0cac5cc7226/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/667c290c56ac5366bd1cc9b7162904c9500d7e17104828be4187f0cac5cc7226/json HTTP/1.1" 200 None ClickHouse node1 started get_instance_ip instance_name=node2 http://localhost:None "GET /v1.46/containers/roottestasyncconnecttomultipleips-gw8-node2-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node2 http://localhost:None "GET /v1.46/containers/roottestasyncconnecttomultipleips-gw8-node2-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node2, ip: 10.5.95.12... http://localhost:None "GET /v1.46/containers/roottestasyncconnecttomultipleips-gw8-node2-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/1db2c4dd28a0f377ae0d5f2a11894d97a15fc127f82cfa0b548cd0dbfd1aa58e/json HTTP/1.1" 200 None ClickHouse node2 started Executing query CREATE TABLE test(t Date, label UInt8) ENGINE = MergeTree PARTITION BY t ORDER BY label; on node1 Executing query INSERT INTO tbl SETTINGS async_insert=true VALUES (3, 'your') on node1 Stderr: Container roottestaggregationmemoryefficient-gw6-node1-1 Stopping Stderr: Container roottestaggregationmemoryefficient-gw6-node2-1 Stopping Stderr: Container roottestaggregationmemoryefficient-gw6-node2-1 Stopped Stderr: Container roottestaggregationmemoryefficient-gw6-node1-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw6/node1/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw6/node1/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw6/node2/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw6/node2/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Skipping quantilesTDigestWeighted Checking groupBitAnd Executing query select hex(initializeAggregation('groupBitAndState', 'foo')) on backward Command:[docker compose --env-file /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw6/.env --project-name roottestaggregationmemoryefficient-gw6 --file /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw6/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw6/node2/docker-compose.yml down --volumes] Skipping groupBitAnd Checking quantileTDigest Executing query select hex(initializeAggregation('quantileTDigestState', 'foo')) on backward Executing query INSERT INTO test SELECT toDate('2022-12-28'), 1; on node1 Executing query INSERT INTO tbl SETTINGS async_insert=true VALUES (4, 'chickens') on node2 Skipping quantileTDigest Checking quantileTDigestWeighted Executing query select hex(initializeAggregation('quantileTDigestWeightedState', 'foo')) on backward Executing query SELECT count(*) FROM test on node1 Skipping quantileTDigestWeighted Checking quantileDeterministic Executing query select hex(initializeAggregation('quantileDeterministicState', 'foo')) on backward run container_id:roottestasyncconnecttomultipleips-gw8-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "echo '2001:3984:3989::1:1118 node1' >> /etc/hosts"] Command:[docker exec roottestasyncconnecttomultipleips-gw8-node2-1 bash -c echo '2001:3984:3989::1:1118 node1' >> /etc/hosts] Skipping quantileDeterministic Checking quantilesTDigest Executing query select hex(initializeAggregation('quantilesTDigestState', 'foo')) on backward Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster' tbl on node1 run container_id:roottestasyncconnecttomultipleips-gw8-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "echo '10.5.95.11 node1' >> /etc/hosts"] Command:[docker exec roottestasyncconnecttomultipleips-gw8-node2-1 bash -c echo '10.5.95.11 node1' >> /etc/hosts] Skipping quantilesTDigest Checking stochasticLogisticRegression Executing query select hex(initializeAggregation('stochasticLogisticRegressionState', 'foo')) on backward Executing query SELECT count(*) from test on node1 Skipping stochasticLogisticRegression Checking argMin Executing query select hex(initializeAggregation('argMinState', 'foo')) on backward Executing query SYSTEM DROP DNS CACHE on node2 Skipping argMin Checking avg Executing query select hex(initializeAggregation('avgState', 'foo')) on backward Executing query BACKUP TABLE tbl ON CLUSTER 'cluster' TO Disk('backups', '17') SETTINGS replica_num=1 on node1 Stderr: Container roottestaggregationmemoryefficient-gw6-node2-1 Stopping Stderr: Container roottestaggregationmemoryefficient-gw6-node1-1 Stopping Stderr: Container roottestaggregationmemoryefficient-gw6-node2-1 Stopped Stderr: Container roottestaggregationmemoryefficient-gw6-node2-1 Removing Stderr: Container roottestaggregationmemoryefficient-gw6-node1-1 Stopped Stderr: Container roottestaggregationmemoryefficient-gw6-node1-1 Removing Stderr: Container roottestaggregationmemoryefficient-gw6-node2-1 Removed Stderr: Container roottestaggregationmemoryefficient-gw6-node1-1 Removed Stderr: Network roottestaggregationmemoryefficient-gw6_default Removing Stderr: Network roottestaggregationmemoryefficient-gw6_default Removed Cleanup called Docker networks for project roottestaggregationmemoryefficient-gw6 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestaggregationmemoryefficient-gw6 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestaggregationmemoryefficient-gw6 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestaggregationmemoryefficient-gw6-.*-1$' --format '{{.ID}}:{{.Names}}'] Executing query SYSTEM DROP DNS CACHE on node1 Stderr: Container roottestbufferprofile-gw7-node_buffer_profile-1 Stopping Stderr: Container roottestbufferprofile-gw7-node_default-1 Stopping Stderr: Container roottestbufferprofile-gw7-node_default-1 Stopped Stderr: Container roottestbufferprofile-gw7-node_buffer_profile-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_buffer_profile/_instances-0-gw7/node_default/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_buffer_profile/_instances-0-gw7/node_default/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Unstopped containers: {} No running containers for project: roottestaggregationmemoryefficient-gw6 Trying to prune unused networks... Command:[bash -c [ -f /ClickHouse/tests/integration/test_buffer_profile/_instances-0-gw7/node_buffer_profile/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_buffer_profile/_instances-0-gw7/node_buffer_profile/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Skipping avg Checking covarSampStable Executing query select hex(initializeAggregation('covarSampStableState', 'foo')) on backward Command:[docker compose --env-file /ClickHouse/tests/integration/test_buffer_profile/_instances-0-gw7/.env --project-name roottestbufferprofile-gw7 --file /ClickHouse/tests/integration/test_buffer_profile/_instances-0-gw7/node_default/docker-compose.yml --file /ClickHouse/tests/integration/test_buffer_profile/_instances-0-gw7/node_buffer_profile/docker-compose.yml down --volumes] Trying to prune unused images... Command:[docker image prune -f] Executing query SELECT count(*) FROM remote('node1', default.test) limit 1; on node2 Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:1 Volumes pruned: 1 test_backward_compatibility/test_cte_distributed.py::test_cte_distributed Running tests in /ClickHouse/tests/integration/test_backward_compatibility/test_cte_distributed.py Cluster start called. is_up=False Executing query DROP TABLE tbl ON CLUSTER 'cluster' SYNC on node1 Skipping covarSampStable Checking quantilesTiming Executing query select hex(initializeAggregation('quantilesTimingState', 'foo')) on backward Docker networks for project roottestbackwardcompatibilityctedistributed-gw6 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestbackwardcompatibilityctedistributed-gw6 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestbackwardcompatibilityctedistributed-gw6 are DRIVER VOLUME NAME Cleanup called Docker networks for project roottestbackwardcompatibilityctedistributed-gw6 are NETWORK ID NAME DRIVER SCOPE Skipping quantilesTiming Checking welchTTest Executing query select hex(initializeAggregation('welchTTestState', 'foo')) on backward Docker containers for project roottestbackwardcompatibilityctedistributed-gw6 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestbackwardcompatibilityctedistributed-gw6 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestbackwardcompatibilityctedistributed-gw6-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestbackwardcompatibilityctedistributed-gw6 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Executing query RESTORE TABLE tbl ON CLUSTER 'cluster' FROM Disk('backups', '17') on node2 Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Skipping welchTTest Checking covarSamp Executing query select hex(initializeAggregation('covarSampState', 'foo')) on backward Stdout:1 Volumes pruned: 1 Setup directory for instance: node1 Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files [] to /ClickHouse/tests/integration/test_backward_compatibility/_instances-cte_distributed-0-gw6/node1/configs/config.d Setup database dir /ClickHouse/tests/integration/test_backward_compatibility/_instances-cte_distributed-0-gw6/node1/database Setup logs dir /ClickHouse/tests/integration/test_backward_compatibility/_instances-cte_distributed-0-gw6/node1/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Setup directory for instance: node2 Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files [] to /ClickHouse/tests/integration/test_backward_compatibility/_instances-cte_distributed-0-gw6/node2/configs/config.d Setup database dir /ClickHouse/tests/integration/test_backward_compatibility/_instances-cte_distributed-0-gw6/node2/database Setup logs dir /ClickHouse/tests/integration/test_backward_compatibility/_instances-cte_distributed-0-gw6/node2/logs Entrypoint cmd: bash -c "trap 'pkill tail' INT TERM; clickhouse server --config-file=/etc/clickhouse-server/config.xml --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log --daemon -- ; coproc tail -f /dev/null; wait $$!" Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw'} stored in /ClickHouse/tests/integration/test_backward_compatibility/_instances-cte_distributed-0-gw6/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_backward_compatibility/_instances-cte_distributed-0-gw6/.env --project-name roottestbackwardcompatibilityctedistributed-gw6 --file /ClickHouse/tests/integration/test_backward_compatibility/_instances-cte_distributed-0-gw6/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_backward_compatibility/_instances-cte_distributed-0-gw6/node2/docker-compose.yml pull] Skipping covarSamp Checking varPopStable Executing query select hex(initializeAggregation('varPopStableState', 'foo')) on backward Skipping varPopStable Checking quantileTiming Executing query select hex(initializeAggregation('quantileTimingState', 'foo')) on backward Skipping quantileTiming Checking leadInFrame Executing query select hex(initializeAggregation('leadInFrameState', 'foo')) on backward Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster' tbl on node1 Skipping leadInFrame Checking quantilesExactExclusive Executing query select hex(initializeAggregation('quantilesExactExclusiveState', 'foo')) on backward Stderr: Container roottestbufferprofile-gw7-node_default-1 Stopping Stderr: Container roottestbufferprofile-gw7-node_buffer_profile-1 Stopping Stderr: Container roottestbufferprofile-gw7-node_default-1 Stopped Stderr: Container roottestbufferprofile-gw7-node_default-1 Removing Stderr: Container roottestbufferprofile-gw7-node_buffer_profile-1 Stopped Stderr: Container roottestbufferprofile-gw7-node_buffer_profile-1 Removing Stderr: Container roottestbufferprofile-gw7-node_default-1 Removed Stderr: Container roottestbufferprofile-gw7-node_buffer_profile-1 Removed Stderr: Network roottestbufferprofile-gw7_default Removing Stderr: Network roottestbufferprofile-gw7_default Removed Cleanup called Docker networks for project roottestbufferprofile-gw7 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestbufferprofile-gw7 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Skipping quantilesExactExclusive Checking maxIntersectionsPosition Executing query select hex(initializeAggregation('maxIntersectionsPositionState', 'foo')) on backward Docker volumes for project roottestbufferprofile-gw7 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestbufferprofile-gw7-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestbufferprofile-gw7 Trying to prune unused networks... Skipping maxIntersectionsPosition Checking quantilesBFloat16 Executing query select hex(initializeAggregation('quantilesBFloat16State', 'foo')) on backward Trying to prune unused images... Command:[docker image prune -f] Executing query SELECT * FROM tbl ORDER BY x on node2 Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost Stdout:1 Volumes pruned: 1 test_cluster_discovery/test_password.py::test_connect_with_password Running tests in /ClickHouse/tests/integration/test_cluster_discovery/test_password.py Cluster start called. is_up=False Docker networks for project roottestclusterdiscoverypassword-gw7 are NETWORK ID NAME DRIVER SCOPE Skipping quantilesBFloat16 Checking quantilesExactInclusive Executing query select hex(initializeAggregation('quantilesExactInclusiveState', 'foo')) on backward Docker containers for project roottestclusterdiscoverypassword-gw7 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Executing query SELECT * FROM tbl ORDER BY x on node1 Docker volumes for project roottestclusterdiscoverypassword-gw7 are DRIVER VOLUME NAME Cleanup called Docker networks for project roottestclusterdiscoverypassword-gw7 are NETWORK ID NAME DRIVER SCOPE Connecting to 172.16.2.3(172.16.2.3):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Docker containers for project roottestclusterdiscoverypassword-gw7 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Skipping quantilesExactInclusive Checking quantileExactInclusive Executing query select hex(initializeAggregation('quantileExactInclusiveState', 'foo')) on backward Docker volumes for project roottestclusterdiscoverypassword-gw7 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestclusterdiscoverypassword-gw7-.*-1$' --format '{{.ID}}:{{.Names}}'] Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw0] PASSED test_backup_restore_on_cluster/test.py::test_replicated_table Unstopped containers: {} No running containers for project: roottestclusterdiscoverypassword-gw7 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Skipping quantileExactInclusive Checking quantileInterpolatedWeighted Executing query select hex(initializeAggregation('quantileInterpolatedWeightedState', 'foo')) on backward Stdout:1 Volumes pruned: 1 Setup directory for instance: node0 Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_cluster_discovery/config/config_with_pwd.xml', '/ClickHouse/tests/integration/test_cluster_discovery/config/config_with_secret1.xml'] to /ClickHouse/tests/integration/test_cluster_discovery/_instances-password-0-gw7/node0/configs/config.d Setup database dir /ClickHouse/tests/integration/test_cluster_discovery/_instances-password-0-gw7/node0/database Setup logs dir /ClickHouse/tests/integration/test_cluster_discovery/_instances-password-0-gw7/node0/logs Entrypoint cmd: bash -c "trap 'pkill tail' INT TERM; clickhouse server --config-file=/etc/clickhouse-server/config.xml --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log --daemon -- ; coproc tail -f /dev/null; wait $$!" Setup directory for instance: node1 Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_cluster_discovery/config/config_with_pwd.xml', '/ClickHouse/tests/integration/test_cluster_discovery/config/config_with_secret2.xml'] to /ClickHouse/tests/integration/test_cluster_discovery/_instances-password-0-gw7/node1/configs/config.d Setup database dir /ClickHouse/tests/integration/test_cluster_discovery/_instances-password-0-gw7/node1/database Setup logs dir /ClickHouse/tests/integration/test_cluster_discovery/_instances-password-0-gw7/node1/logs Entrypoint cmd: bash -c "trap 'pkill tail' INT TERM; clickhouse server --config-file=/etc/clickhouse-server/config.xml --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log --daemon -- ; coproc tail -f /dev/null; wait $$!" Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'keeper_binary': '/clickhouse', 'keeper_cmd_prefix': 'clickhouse keeper', 'image': 'altinityinfra/integration-test:8b2301119731', 'user': '0', 'keeper_fs': 'bind', 'keeper_logs_dir1': '/ClickHouse/tests/integration/test_cluster_discovery/_instances-password-0-gw7/keeper1/log', 'keeper_config_dir1': '/ClickHouse/tests/integration/test_cluster_discovery/_instances-password-0-gw7/keeper1/config', 'keeper_db_dir1': '/ClickHouse/tests/integration/test_cluster_discovery/_instances-password-0-gw7/keeper1/coordination', 'keeper_logs_dir2': '/ClickHouse/tests/integration/test_cluster_discovery/_instances-password-0-gw7/keeper2/log', 'keeper_config_dir2': '/ClickHouse/tests/integration/test_cluster_discovery/_instances-password-0-gw7/keeper2/config', 'keeper_db_dir2': '/ClickHouse/tests/integration/test_cluster_discovery/_instances-password-0-gw7/keeper2/coordination', 'keeper_logs_dir3': '/ClickHouse/tests/integration/test_cluster_discovery/_instances-password-0-gw7/keeper3/log', 'keeper_config_dir3': '/ClickHouse/tests/integration/test_cluster_discovery/_instances-password-0-gw7/keeper3/config', 'keeper_db_dir3': '/ClickHouse/tests/integration/test_cluster_discovery/_instances-password-0-gw7/keeper3/coordination'} stored in /ClickHouse/tests/integration/test_cluster_discovery/_instances-password-0-gw7/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_cluster_discovery/_instances-password-0-gw7/.env --project-name roottestclusterdiscoverypassword-gw7 --file /ClickHouse/tests/integration/test_cluster_discovery/_instances-password-0-gw7/node0/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_cluster_discovery/_instances-password-0-gw7/node1/docker-compose.yml pull] Connection dropped: outstanding heartbeat ping not received Transition to CONNECTING Zookeeper connection lost Skipping quantileInterpolatedWeighted Checking quantileExactHigh Executing query select hex(initializeAggregation('quantileExactHighState', 'foo')) on backward Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Connecting to 172.16.2.3(172.16.2.3):2181, use_ssl: False Skipping quantileExactHigh Checking groupBitmap Executing query select hex(initializeAggregation('groupBitmapState', 'foo')) on backward Connecting to 172.16.2.3(172.16.2.3):2181, use_ssl: False Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Skipping groupBitmap Checking quantilesInterpolatedWeighted Executing query select hex(initializeAggregation('quantilesInterpolatedWeightedState', 'foo')) on backward Skipping quantilesInterpolatedWeighted Checking quantile Executing query select hex(initializeAggregation('quantileState', 'foo')) on backward Skipping quantile Checking groupArraySample Executing query select hex(initializeAggregation('groupArraySampleState', 'foo')) on backward Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Skipping groupArraySample Checking quantilesExactLow Executing query select hex(initializeAggregation('quantilesExactLowState', 'foo')) on backward Skipping quantilesExactLow Checking groupUniqArray Executing query select hex(initializeAggregation('groupUniqArrayState', 'foo')) on backward Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Executing query select hex(initializeAggregation('groupUniqArrayState', 'foo')) on upstream test_backup_restore_on_cluster/test.py::test_replicated_table_after_alters Executing query CREATE TABLE tbl ON CLUSTER 'cluster' (x Int32) ENGINE=ReplicatedMergeTree('/clickhouse/tables/tbl/', '{replica}')ORDER BY x on node1 OK groupUniqArray Checking sumWithOverflow Executing query select hex(initializeAggregation('sumWithOverflowState', 'foo')) on backward Executing query INSERT INTO tbl VALUES (1) on node1 Skipping sumWithOverflow Checking sequenceCount Executing query select hex(initializeAggregation('sequenceCountState', 'foo')) on backward Executing query ALTER TABLE tbl ADD COLUMN y Int32 on node1 Skipping sequenceCount Checking quantilesDeterministic Executing query select hex(initializeAggregation('quantilesDeterministicState', 'foo')) on backward Executing query INSERT INTO tbl VALUES (2, 20) on node1 Skipping quantilesDeterministic Checking quantilesExactHigh Executing query select hex(initializeAggregation('quantilesExactHighState', 'foo')) on backward Stderr: Container roottestalternativekeeperconfig-gw2-node1-1 Stopping Stderr: Container roottestalternativekeeperconfig-gw2-node3-1 Stopping Stderr: Container roottestalternativekeeperconfig-gw2-node2-1 Stopping Stderr: Container roottestalternativekeeperconfig-gw2-node2-1 Stopped Stderr: Container roottestalternativekeeperconfig-gw2-node3-1 Stopped Stderr: Container roottestalternativekeeperconfig-gw2-node1-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_alternative_keeper_config/_instances-0-gw2/node1/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_alternative_keeper_config/_instances-0-gw2/node1/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_alternative_keeper_config/_instances-0-gw2/node2/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_alternative_keeper_config/_instances-0-gw2/node2/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Executing query ALTER TABLE tbl ADD COLUMN z Int32 on node1 Command:[bash -c [ -f /ClickHouse/tests/integration/test_alternative_keeper_config/_instances-0-gw2/node3/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_alternative_keeper_config/_instances-0-gw2/node3/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Skipping quantilesExactHigh Checking sumCount Executing query select hex(initializeAggregation('sumCountState', 'foo')) on backward Command:[docker compose --env-file /ClickHouse/tests/integration/test_alternative_keeper_config/_instances-0-gw2/.env --project-name roottestalternativekeeperconfig-gw2 --file /ClickHouse/tests/integration/test_alternative_keeper_config/_instances-0-gw2/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_alternative_keeper_config/_instances-0-gw2/node2/docker-compose.yml --file /ClickHouse/tests/integration/test_alternative_keeper_config/_instances-0-gw2/node3/docker-compose.yml down --volumes] Executing query INSERT INTO tbl VALUES (3, 30, 300) on node1 Skipping sumCount Checking groupArrayInsertAt Executing query select hex(initializeAggregation('groupArrayInsertAtState', 'foo')) on backward Executing query BACKUP TABLE tbl ON CLUSTER 'cluster' TO Disk('backups', '18') on node1 Skipping groupArrayInsertAt Checking quantileExact Executing query select hex(initializeAggregation('quantileExactState', 'foo')) on backward Skipping quantileExact Checking anyHeavy Executing query select hex(initializeAggregation('anyHeavyState', 'foo')) on backward Executing query select hex(initializeAggregation('anyHeavyState', 'foo')) on upstream Executing query DROP TABLE tbl ON CLUSTER 'cluster' SYNC on node1 Executing query select finalizeAggregation(unhex('04000000666F6F000100000000000000')::AggregateFunction(anyHeavy, String)) on backward Executing query select finalizeAggregation(unhex('04000000666F6F000000000000000000')::AggregateFunction(anyHeavy, String)) on upstream Executing query RESTORE TABLE tbl ON CLUSTER 'cluster' FROM Disk('backups', '18') on node1 OK anyHeavy (but different intermediate states) Checking theilsU Executing query select hex(initializeAggregation('theilsUState', 'foo')) on backward Stderr: Container roottestalternativekeeperconfig-gw2-node3-1 Stopping Stderr: Container roottestalternativekeeperconfig-gw2-node2-1 Stopping Stderr: Container roottestalternativekeeperconfig-gw2-node1-1 Stopping Stderr: Container roottestalternativekeeperconfig-gw2-node3-1 Stopped Stderr: Container roottestalternativekeeperconfig-gw2-node3-1 Removing Stderr: Container roottestalternativekeeperconfig-gw2-node2-1 Stopped Stderr: Container roottestalternativekeeperconfig-gw2-node2-1 Removing Stderr: Container roottestalternativekeeperconfig-gw2-node1-1 Stopped Stderr: Container roottestalternativekeeperconfig-gw2-node1-1 Removing Stderr: Container roottestalternativekeeperconfig-gw2-node2-1 Removed Stderr: Container roottestalternativekeeperconfig-gw2-node3-1 Removed Stderr: Container roottestalternativekeeperconfig-gw2-node1-1 Removed Stderr: Network roottestalternativekeeperconfig-gw2_default Removing Stderr: Network roottestalternativekeeperconfig-gw2_default Removed Cleanup called Docker networks for project roottestalternativekeeperconfig-gw2 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestalternativekeeperconfig-gw2 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestalternativekeeperconfig-gw2 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestalternativekeeperconfig-gw2-.*-1$' --format '{{.ID}}:{{.Names}}'] Skipping theilsU Checking maxMappedArrays Executing query select hex(initializeAggregation('maxMappedArraysState', 'foo')) on backward Unstopped containers: {} No running containers for project: roottestalternativekeeperconfig-gw2 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Skipping maxMappedArrays Checking stddevPop Executing query select hex(initializeAggregation('stddevPopState', 'foo')) on backward Stdout:1 Volumes pruned: 1 test_backup_restore_new/test_shutdown_wait_backup.py::test_shutdown_wait_backup Running tests in /ClickHouse/tests/integration/test_backup_restore_new/test_shutdown_wait_backup.py Cluster start called. is_up=False Docker networks for project roottestbackuprestorenewshutdownwaitbackup-gw2 are NETWORK ID NAME DRIVER SCOPE Executing query OPTIMIZE TABLE tbl FINAL on node1 Skipping stddevPop Checking groupArrayLast Docker containers for project roottestbackuprestorenewshutdownwaitbackup-gw2 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Executing query select hex(initializeAggregation('groupArrayLastState', 'foo')) on backward Docker volumes for project roottestbackuprestorenewshutdownwaitbackup-gw2 are DRIVER VOLUME NAME Cleanup called Docker networks for project roottestbackuprestorenewshutdownwaitbackup-gw2 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestbackuprestorenewshutdownwaitbackup-gw2 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestbackuprestorenewshutdownwaitbackup-gw2 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestbackuprestorenewshutdownwaitbackup-gw2-.*-1$' --format '{{.ID}}:{{.Names}}'] Executing query SELECT * FROM tbl ORDER BY x on node1 Skipping groupArrayLast Checking groupArray Executing query select hex(initializeAggregation('groupArrayState', 'foo')) on backward Unstopped containers: {} No running containers for project: roottestbackuprestorenewshutdownwaitbackup-gw2 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Executing query SELECT * FROM tbl ORDER BY x on node2 Executing query select hex(initializeAggregation('groupArrayState', 'foo')) on upstream Stdout:1 Volumes pruned: 1 Setup directory for instance: node Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_backup_restore_new/configs/backups_disk.xml', '/ClickHouse/tests/integration/test_backup_restore_new/configs/slow_backups.xml'] to /ClickHouse/tests/integration/test_backup_restore_new/_instances-shutdown_wait_backup-0-gw2/node/configs/config.d Setup database dir /ClickHouse/tests/integration/test_backup_restore_new/_instances-shutdown_wait_backup-0-gw2/node/database Setup logs dir /ClickHouse/tests/integration/test_backup_restore_new/_instances-shutdown_wait_backup-0-gw2/node/logs Entrypoint cmd: bash -c "trap 'pkill tail' INT TERM; clickhouse server --config-file=/etc/clickhouse-server/config.xml --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log --daemon -- ; coproc tail -f /dev/null; wait $$!" external_dir_abs_path=/ClickHouse/tests/integration/test_backup_restore_new/_instances-shutdown_wait_backup-0-gw2/backups Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw'} stored in /ClickHouse/tests/integration/test_backup_restore_new/_instances-shutdown_wait_backup-0-gw2/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_backup_restore_new/_instances-shutdown_wait_backup-0-gw2/.env --project-name roottestbackuprestorenewshutdownwaitbackup-gw2 --file /ClickHouse/tests/integration/test_backup_restore_new/_instances-shutdown_wait_backup-0-gw2/node/docker-compose.yml pull] OK groupArray Checking first_value Executing query select hex(initializeAggregation('first_valueState', 'foo')) on backward Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw0] PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_after_alters Executing query select hex(initializeAggregation('first_valueState', 'foo')) on upstream OK first_value Checking exponentialTimeDecayedAvg Executing query select hex(initializeAggregation('exponentialTimeDecayedAvgState', 'foo')) on backward Skipping exponentialTimeDecayedAvg Checking quantileExactExclusive Executing query select hex(initializeAggregation('quantileExactExclusiveState', 'foo')) on backward Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Skipping quantileExactExclusive Checking count Executing query select hex(initializeAggregation('countState', 'foo')) on backward Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Executing query select hex(initializeAggregation('countState', 'foo')) on upstream OK count Checking stddevPopStable Executing query select hex(initializeAggregation('stddevPopStableState', 'foo')) on backward Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Skipping stddevPopStable Checking quantileExactLow Executing query select hex(initializeAggregation('quantileExactLowState', 'foo')) on backward Skipping quantileExactLow Checking avgWeighted Executing query select hex(initializeAggregation('avgWeightedState', 'foo')) on backward Command:[docker compose --env-file /ClickHouse/tests/integration/test_async_connect_to_multiple_ips/_instances-0-gw8/.env --project-name roottestasyncconnecttomultipleips-gw8 --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_net.yml --file /ClickHouse/tests/integration/test_async_connect_to_multiple_ips/_instances-0-gw8/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_async_connect_to_multiple_ips/_instances-0-gw8/node2/docker-compose.yml stop --timeout 20] [gw8] PASSED test_async_connect_to_multiple_ips/test.py::test Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Skipping avgWeighted Checking analysisOfVariance Executing query select hex(initializeAggregation('analysisOfVarianceState', 'foo')) on backward Skipping analysisOfVariance Checking groupBitAnd Executing query select hex(initializeAggregation('groupBitAndState', 'foo')) on backward Skipping groupBitAnd Checking varSamp Executing query select hex(initializeAggregation('varSampState', 'foo')) on backward test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_bigger_cluster Executing query CREATE TABLE tbl ON CLUSTER 'cluster' (x UInt32) ENGINE=ReplicatedMergeTree('/clickhouse/tables/tbl/', '{replica}')ORDER BY x on node1 Skipping varSamp Checking quantileTimingWeighted Executing query select hex(initializeAggregation('quantileTimingWeightedState', 'foo')) on backward Skipping quantileTimingWeighted Checking covarSamp Executing query select hex(initializeAggregation('covarSampState', 'foo')) on backward Executing query INSERT INTO tbl VALUES (111) on node1 Skipping covarSamp Checking varPop Executing query select hex(initializeAggregation('varPopState', 'foo')) on backward Executing query INSERT INTO tbl VALUES (222) on node2 Skipping varPop Checking quantileBFloat16 Executing query select hex(initializeAggregation('quantileBFloat16State', 'foo')) on backward Executing query BACKUP TABLE tbl ON CLUSTER 'cluster' TO Disk('backups', '19') on node1 Skipping quantileBFloat16 Checking quantileTDigest Executing query select hex(initializeAggregation('quantileTDigestState', 'foo')) on backward Skipping quantileTDigest Checking quantileTiming Executing query select hex(initializeAggregation('quantileTimingState', 'foo')) on backward Executing query DROP TABLE tbl ON CLUSTER 'cluster' SYNC on node1 Skipping quantileTiming Checking quantileExactHigh Executing query select hex(initializeAggregation('quantileExactHighState', 'foo')) on backward Skipping quantileExactHigh Checking quantileInterpolatedWeighted Executing query select hex(initializeAggregation('quantileInterpolatedWeightedState', 'foo')) on backward Skipping quantileInterpolatedWeighted Checking groupBitXor Executing query select hex(initializeAggregation('groupBitXorState', 'foo')) on backward Executing query RESTORE TABLE tbl ON CLUSTER 'cluster3' FROM Disk('backups', '19') on node1 Skipping groupBitXor Checking quantile Executing query select hex(initializeAggregation('quantileState', 'foo')) on backward Skipping quantile Checking covarPop Executing query select hex(initializeAggregation('covarPopState', 'foo')) on backward Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster3' tbl on node1 Skipping covarPop Checking quantileBFloat16Weighted Executing query select hex(initializeAggregation('quantileBFloat16WeightedState', 'foo')) on backward Skipping quantileBFloat16Weighted Checking quantileExactWeighted Executing query select hex(initializeAggregation('quantileExactWeightedState', 'foo')) on backward Skipping quantileExactWeighted Checking quantileTDigestWeighted Executing query select hex(initializeAggregation('quantileTDigestWeightedState', 'foo')) on backward Skipping quantileTDigestWeighted Checking quantileExact Executing query select hex(initializeAggregation('quantileExactState', 'foo')) on backward Skipping quantileExact Checking stddevPop Executing query select hex(initializeAggregation('stddevPopState', 'foo')) on backward Skipping stddevPop Checking stddevSamp Executing query select hex(initializeAggregation('stddevSampState', 'foo')) on backward Skipping stddevSamp Checking quantileExactLow Executing query select hex(initializeAggregation('quantileExactLowState', 'foo')) on backward Executing query SELECT * FROM tbl ORDER BY x on node1 Skipping quantileExactLow Checking groupBitOr Executing query select hex(initializeAggregation('groupBitOrState', 'foo')) on backward Executing query SELECT * FROM tbl ORDER BY x on node2 Skipping groupBitOr Checking quantileDeterministic Executing query select hex(initializeAggregation('quantileDeterministicState', 'foo')) on backward Skipping quantileDeterministic Aggregate functions: 163, Failed: 0, skipped: 142, passed: 21 [gw9] PASSED test_backward_compatibility/test_functions.py::test_aggregate_states test_backward_compatibility/test_functions.py::test_string_functions Executing query SELECT value FROM system.build_options WHERE name = 'CXX_FLAGS' on upstream Executing query SELECT * FROM tbl ORDER BY x on node3 Executing query SELECT value FROM system.build_options WHERE name = 'CXX_FLAGS' on upstream Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw0] PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_bigger_cluster Executing query SELECT value FROM system.build_options WHERE name = 'CXX_FLAGS' on upstream Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Executing query SELECT if(NOT empty(alias_to), alias_to, name) FROM system.functions WHERE is_aggregate = 0 on backward Got 1172 functions Checking polygonsEqualsCartesian('foo') Executing query select polygonsEqualsCartesian('foo') on backward Skipping polygonsEqualsCartesian Checking h3GetPentagonIndexes('foo') Executing query select h3GetPentagonIndexes('foo') on backward Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Skipping h3GetPentagonIndexes Checking JSON_VALUE('foo') Executing query select JSON_VALUE('foo') on backward Skipping JSON_VALUE Checking intHash64('foo') Executing query select intHash64('foo') on backward Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Connection dropped: socket connection error: No route to host Connection dropped: socket connection error: No route to host Skipping intHash64 Checking intHash32('foo') Executing query select intHash32('foo') on backward Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Skipping intHash32 Checking multiSearchFirstIndexUTF8('foo') Executing query select multiSearchFirstIndexUTF8('foo') on backward Skipping multiSearchFirstIndexUTF8 Checking getSetting('foo') Executing query select getSetting('foo') on backward test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_smaller_cluster Executing query CREATE TABLE tbl ON CLUSTER 'cluster' (x UInt32) ENGINE=ReplicatedMergeTree('/clickhouse/tables/tbl/', '{replica}')ORDER BY x on node1 Connecting to 172.16.2.3(172.16.2.3):2181, use_ssl: False Skipping getSetting Checking isValidUTF8('foo') Executing query select isValidUTF8('foo') on backward Connecting to 172.16.2.3(172.16.2.3):2181, use_ssl: False Executing query select isValidUTF8('foo') on upstream Executing query INSERT INTO tbl VALUES (111) on node1 OK isValidUTF8 Checking toIntervalHour('foo') Executing query select toIntervalHour('foo') on backward Executing query INSERT INTO tbl VALUES (222) on node2 Skipping toIntervalHour Checking toIntervalMinute('foo') Executing query select toIntervalMinute('foo') on backward Executing query BACKUP TABLE tbl ON CLUSTER 'cluster' TO Disk('backups', '20') on node1 Skipping toIntervalMinute Checking toIntervalMillisecond('foo') Executing query select toIntervalMillisecond('foo') on backward Skipping toIntervalMillisecond Checking parseDateTime64BestEffortUSOrZero('foo') Executing query select parseDateTime64BestEffortUSOrZero('foo') on backward Executing query DROP TABLE tbl ON CLUSTER 'cluster' SYNC on node1 Executing query select parseDateTime64BestEffortUSOrZero('foo') on upstream OK parseDateTime64BestEffortUSOrZero Checking parseDateTime64BestEffortOrZero('foo') Executing query select parseDateTime64BestEffortOrZero('foo') on backward Executing query RESTORE TABLE tbl ON CLUSTER 'cluster1' FROM Disk('backups', '20') on node1 Executing query select parseDateTime64BestEffortOrZero('foo') on upstream OK parseDateTime64BestEffortOrZero Checking parseDateTime64BestEffort('foo') Executing query select parseDateTime64BestEffort('foo') on backward Stderr: zoo2 Skipped - Image is already being pulled by zoo1 Stderr: zoo3 Skipped - Image is already being pulled by zoo1 Stderr: node Pulling Stderr: zoo1 Pulling Stderr: zoo1 Pulled Stderr: node Pulled Setup ZooKeeper Creating internal ZooKeeper dirs: ['/ClickHouse/tests/integration/test_backward_compatibility/_instances-convert_ordinary-0-gw5/keeper1/log', '/ClickHouse/tests/integration/test_backward_compatibility/_instances-convert_ordinary-0-gw5/keeper1/config', '/ClickHouse/tests/integration/test_backward_compatibility/_instances-convert_ordinary-0-gw5/keeper1/coordination', '/ClickHouse/tests/integration/test_backward_compatibility/_instances-convert_ordinary-0-gw5/keeper2/log', '/ClickHouse/tests/integration/test_backward_compatibility/_instances-convert_ordinary-0-gw5/keeper2/config', '/ClickHouse/tests/integration/test_backward_compatibility/_instances-convert_ordinary-0-gw5/keeper2/coordination', '/ClickHouse/tests/integration/test_backward_compatibility/_instances-convert_ordinary-0-gw5/keeper3/log', '/ClickHouse/tests/integration/test_backward_compatibility/_instances-convert_ordinary-0-gw5/keeper3/config', '/ClickHouse/tests/integration/test_backward_compatibility/_instances-convert_ordinary-0-gw5/keeper3/coordination'] Command:[docker compose --project-name roottestbackwardcompatibilityconvertordinary-gw5 --env-file /ClickHouse/tests/integration/test_backward_compatibility/_instances-convert_ordinary-0-gw5/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --verbose up -d] Skipping parseDateTime64BestEffort Checking parseDateTime32BestEffort('foo') Executing query select parseDateTime32BestEffort('foo') on backward Executing query SELECT * FROM tbl ORDER BY x on node1 Skipping parseDateTime32BestEffort Checking parseDateTimeBestEffortOrNull('foo') Executing query select parseDateTimeBestEffortOrNull('foo') on backward Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw0] PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_smaller_cluster Stderr: node1 Pulling Stderr: node2 Pulling Stderr: node2 Pulled Stderr: node1 Pulled ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_backward_compatibility/_instances-cte_distributed-0-gw6/.env --project-name roottestbackwardcompatibilityctedistributed-gw6 --file /ClickHouse/tests/integration/test_backward_compatibility/_instances-cte_distributed-0-gw6/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_backward_compatibility/_instances-cte_distributed-0-gw6/node2/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_backward_compatibility/_instances-cte_distributed-0-gw6/.env --project-name roottestbackwardcompatibilityctedistributed-gw6 --file /ClickHouse/tests/integration/test_backward_compatibility/_instances-cte_distributed-0-gw6/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_backward_compatibility/_instances-cte_distributed-0-gw6/node2/docker-compose.yml up -d --no-recreate] Stderr: node Pulling Stderr: node Pulled ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_backup_restore_new/_instances-shutdown_wait_backup-0-gw2/.env --project-name roottestbackuprestorenewshutdownwaitbackup-gw2 --file /ClickHouse/tests/integration/test_backup_restore_new/_instances-shutdown_wait_backup-0-gw2/node/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_backup_restore_new/_instances-shutdown_wait_backup-0-gw2/.env --project-name roottestbackuprestorenewshutdownwaitbackup-gw2 --file /ClickHouse/tests/integration/test_backup_restore_new/_instances-shutdown_wait_backup-0-gw2/node/docker-compose.yml up -d --no-recreate] Executing query select parseDateTimeBestEffortOrNull('foo') on upstream OK parseDateTimeBestEffortOrNull Checking parseDateTimeBestEffortOrZero('foo') Executing query select parseDateTimeBestEffortOrZero('foo') on backward Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select parseDateTimeBestEffortOrZero('foo') on upstream Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 OK parseDateTimeBestEffortOrZero Checking toIPv6OrNull('foo') Executing query select toIPv6OrNull('foo') on backward Stderr: zoo1 Skipped - Image is already being pulled by zoo3 Stderr: node1 Skipped - Image is already being pulled by zoo3 Stderr: node0 Skipped - Image is already being pulled by zoo3 Stderr: zoo2 Skipped - Image is already being pulled by zoo3 Stderr: zoo3 Pulling Stderr: zoo3 Pulled Setup ZooKeeper Creating internal ZooKeeper dirs: ['/ClickHouse/tests/integration/test_cluster_discovery/_instances-password-0-gw7/keeper1/log', '/ClickHouse/tests/integration/test_cluster_discovery/_instances-password-0-gw7/keeper1/config', '/ClickHouse/tests/integration/test_cluster_discovery/_instances-password-0-gw7/keeper1/coordination', '/ClickHouse/tests/integration/test_cluster_discovery/_instances-password-0-gw7/keeper2/log', '/ClickHouse/tests/integration/test_cluster_discovery/_instances-password-0-gw7/keeper2/config', '/ClickHouse/tests/integration/test_cluster_discovery/_instances-password-0-gw7/keeper2/coordination', '/ClickHouse/tests/integration/test_cluster_discovery/_instances-password-0-gw7/keeper3/log', '/ClickHouse/tests/integration/test_cluster_discovery/_instances-password-0-gw7/keeper3/config', '/ClickHouse/tests/integration/test_cluster_discovery/_instances-password-0-gw7/keeper3/coordination'] Command:[docker compose --project-name roottestclusterdiscoverypassword-gw7 --env-file /ClickHouse/tests/integration/test_cluster_discovery/_instances-password-0-gw7/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --verbose up -d] Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select toIPv6OrNull('foo') on upstream Stderr:time="2025-04-02T03:34:43Z" level=trace msg="Docker Desktop integration not enabled" Stderr: Network roottestbackwardcompatibilityconvertordinary-gw5_default Creating Stderr: Network roottestbackwardcompatibilityconvertordinary-gw5_default Created Stderr: Container roottestbackwardcompatibilityconvertordinary-gw5-zoo3-1 Creating Stderr: Container roottestbackwardcompatibilityconvertordinary-gw5-zoo1-1 Creating Stderr: Container roottestbackwardcompatibilityconvertordinary-gw5-zoo2-1 Creating Stderr: Container roottestbackwardcompatibilityconvertordinary-gw5-zoo2-1 Created Stderr: Container roottestbackwardcompatibilityconvertordinary-gw5-zoo3-1 Created Stderr: Container roottestbackwardcompatibilityconvertordinary-gw5-zoo1-1 Created Stderr: Container roottestbackwardcompatibilityconvertordinary-gw5-zoo3-1 Starting Stderr: Container roottestbackwardcompatibilityconvertordinary-gw5-zoo2-1 Starting Stderr: Container roottestbackwardcompatibilityconvertordinary-gw5-zoo1-1 Starting Stderr: Container roottestbackwardcompatibilityconvertordinary-gw5-zoo2-1 Started Stderr: Container roottestbackwardcompatibilityconvertordinary-gw5-zoo1-1 Started Stderr: Container roottestbackwardcompatibilityconvertordinary-gw5-zoo3-1 Started Stderr:time="2025-04-02T03:34:44Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T03:34:44Z" level=debug msg="otel error" error="" Wait ZooKeeper to start get_instance_ip instance_name=zoo1 http://localhost:None "GET /v1.46/containers/roottestbackwardcompatibilityconvertordinary-gw5-zoo1-1/json HTTP/1.1" 200 None get_kazoo_client: zoo1, ip:172.16.1.3, port:2181, use_ssl:False Connecting to 172.16.1.3(172.16.1.3):2181, use_ssl: False Connection dropped: socket connection error: Connection refused OK toIPv6OrNull Checking toIPv4OrNull('foo') Executing query select toIPv4OrNull('foo') on backward Connecting to 172.16.1.3(172.16.1.3):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Executing query select toIPv4OrNull('foo') on upstream Connecting to 172.16.1.3(172.16.1.3):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Stderr: Network roottestbackuprestorenewshutdownwaitbackup-gw2_default Creating Stderr: Network roottestbackuprestorenewshutdownwaitbackup-gw2_default Created Stderr: Container roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 Creating Stderr: Container roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 Created Stderr: Container roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 Starting Stderr: Container roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 Started ClickHouse instance created get_instance_ip instance_name=node http://localhost:None "GET /v1.46/containers/roottestbackuprestorenewshutdownwaitbackup-gw2-node-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node http://localhost:None "GET /v1.46/containers/roottestbackuprestorenewshutdownwaitbackup-gw2-node-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node, ip: 172.16.4.2... http://localhost:None "GET /v1.46/containers/roottestbackuprestorenewshutdownwaitbackup-gw2-node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/a15792276f654be0ad41b5bb01fc2b6028bfa5ac9cbd191917c97c7b902fbf47/json HTTP/1.1" 200 None OK toIPv4OrNull Checking toUUIDOrNull('foo') Executing query select toUUIDOrNull('foo') on backward Executing query select toUUIDOrNull('foo') on upstream test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_def Connecting to 172.16.1.3(172.16.1.3):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Executing query CREATE TABLE tbl (x UInt8, y String) ENGINE=ReplicatedMergeTree('/clickhouse/tables/tbl/', '{replica}')ORDER BY tuple() on node1 http://localhost:None "GET /v1.46/containers/a15792276f654be0ad41b5bb01fc2b6028bfa5ac9cbd191917c97c7b902fbf47/json HTTP/1.1" 200 None Stderr: Network roottestbackwardcompatibilityctedistributed-gw6_default Creating Stderr: Network roottestbackwardcompatibilityctedistributed-gw6_default Created Stderr: Container roottestbackwardcompatibilityctedistributed-gw6-node1-1 Creating Stderr: Container roottestbackwardcompatibilityctedistributed-gw6-node2-1 Creating Stderr: Container roottestbackwardcompatibilityctedistributed-gw6-node2-1 Created Stderr: Container roottestbackwardcompatibilityctedistributed-gw6-node1-1 Created Stderr: Container roottestbackwardcompatibilityctedistributed-gw6-node1-1 Starting Stderr: Container roottestbackwardcompatibilityctedistributed-gw6-node2-1 Starting Stderr: Container roottestbackwardcompatibilityctedistributed-gw6-node1-1 Started Stderr: Container roottestbackwardcompatibilityctedistributed-gw6-node2-1 Started ClickHouse instance created get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottestbackwardcompatibilityctedistributed-gw6-node1-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottestbackwardcompatibilityctedistributed-gw6-node1-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node1, ip: 172.16.6.2... http://localhost:None "GET /v1.46/containers/roottestbackwardcompatibilityctedistributed-gw6-node1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/90a7a5daf61537f8456022dbe173ad70446fd5dd70458eb1bd2984a977a12042/json HTTP/1.1" 200 None OK toUUIDOrNull Checking parseDateTimeBestEffort('foo') Executing query select parseDateTimeBestEffort('foo') on backward Executing query CREATE TABLE tbl (x UInt8, y String) ENGINE=ReplicatedMergeTree('/clickhouse/tables/tbl/', '{replica}')ORDER BY tuple() on node2 http://localhost:None "GET /v1.46/containers/a15792276f654be0ad41b5bb01fc2b6028bfa5ac9cbd191917c97c7b902fbf47/json HTTP/1.1" 200 None Skipping parseDateTimeBestEffort Checking toDecimal128OrNull('foo') Executing query select toDecimal128OrNull('foo') on backward http://localhost:None "GET /v1.46/containers/90a7a5daf61537f8456022dbe173ad70446fd5dd70458eb1bd2984a977a12042/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/a15792276f654be0ad41b5bb01fc2b6028bfa5ac9cbd191917c97c7b902fbf47/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/90a7a5daf61537f8456022dbe173ad70446fd5dd70458eb1bd2984a977a12042/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/a15792276f654be0ad41b5bb01fc2b6028bfa5ac9cbd191917c97c7b902fbf47/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/90a7a5daf61537f8456022dbe173ad70446fd5dd70458eb1bd2984a977a12042/json HTTP/1.1" 200 None ClickHouse node1 started get_instance_ip instance_name=node2 http://localhost:None "GET /v1.46/containers/roottestbackwardcompatibilityctedistributed-gw6-node2-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node2 http://localhost:None "GET /v1.46/containers/roottestbackwardcompatibilityctedistributed-gw6-node2-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node2, ip: 172.16.6.3... http://localhost:None "GET /v1.46/containers/roottestbackwardcompatibilityctedistributed-gw6-node2-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/1a3e090571812d86f699506755aaef001ef74a01aac15443c854bd39a62dc192/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/a15792276f654be0ad41b5bb01fc2b6028bfa5ac9cbd191917c97c7b902fbf47/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/1a3e090571812d86f699506755aaef001ef74a01aac15443c854bd39a62dc192/json HTTP/1.1" 200 None Skipping toDecimal128OrNull Checking toDecimal64OrNull('foo') Executing query select toDecimal64OrNull('foo') on backward http://localhost:None "GET /v1.46/containers/a15792276f654be0ad41b5bb01fc2b6028bfa5ac9cbd191917c97c7b902fbf47/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/1a3e090571812d86f699506755aaef001ef74a01aac15443c854bd39a62dc192/json HTTP/1.1" 200 None ClickHouse node2 started Executing query WITH quantile(0.05)(cnt) as p05, quantile(0.95)(cnt) as p95, p95 - p05 as inter_percentile_range SELECT sum(cnt) as total_requests, count() as data_points, inter_percentile_range FROM ( SELECT count() as cnt FROM remote('node{1,2}', numbers(10)) GROUP BY number ) on node2 Executing query SYSTEM STOP REPLICATION QUEUES tbl on node2 http://localhost:None "GET /v1.46/containers/a15792276f654be0ad41b5bb01fc2b6028bfa5ac9cbd191917c97c7b902fbf47/json HTTP/1.1" 200 None ClickHouse node started Executing query CREATE TABLE tbl (x UInt64) ENGINE=MergeTree() ORDER BY tuple() PARTITION BY x%5 on node Stderr: Container roottestcompatibilitymergetreesettings-gw3-node3-1 Stopping Stderr: Container roottestcompatibilitymergetreesettings-gw3-node1-1 Stopping Stderr: Container roottestcompatibilitymergetreesettings-gw3-node2-1 Stopping Stderr: Container roottestcompatibilitymergetreesettings-gw3-node3-1 Stopped Stderr: Container roottestcompatibilitymergetreesettings-gw3-node2-1 Stopped Stderr: Container roottestcompatibilitymergetreesettings-gw3-node1-1 Stopped Stderr: Container roottestcompatibilitymergetreesettings-gw3-zoo2-1 Stopping Stderr: Container roottestcompatibilitymergetreesettings-gw3-zoo1-1 Stopping Stderr: Container roottestcompatibilitymergetreesettings-gw3-zoo3-1 Stopping Stderr: Container roottestcompatibilitymergetreesettings-gw3-zoo1-1 Stopped Stderr: Container roottestcompatibilitymergetreesettings-gw3-zoo3-1 Stopped Stderr: Container roottestcompatibilitymergetreesettings-gw3-zoo2-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_compatibility_merge_tree_settings/_instances-0-gw3/node1/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_compatibility_merge_tree_settings/_instances-0-gw3/node1/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_compatibility_merge_tree_settings/_instances-0-gw3/node2/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_compatibility_merge_tree_settings/_instances-0-gw3/node2/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_compatibility_merge_tree_settings/_instances-0-gw3/node3/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_compatibility_merge_tree_settings/_instances-0-gw3/node3/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Stderr:time="2025-04-02T03:34:44Z" level=trace msg="Docker Desktop integration not enabled" Stderr: Network roottestclusterdiscoverypassword-gw7_default Creating Stderr: Network roottestclusterdiscoverypassword-gw7_default Created Stderr: Container roottestclusterdiscoverypassword-gw7-zoo2-1 Creating Stderr: Container roottestclusterdiscoverypassword-gw7-zoo3-1 Creating Stderr: Container roottestclusterdiscoverypassword-gw7-zoo1-1 Creating Stderr: Container roottestclusterdiscoverypassword-gw7-zoo2-1 Created Stderr: Container roottestclusterdiscoverypassword-gw7-zoo3-1 Created Stderr: Container roottestclusterdiscoverypassword-gw7-zoo1-1 Created Stderr: Container roottestclusterdiscoverypassword-gw7-zoo2-1 Starting Stderr: Container roottestclusterdiscoverypassword-gw7-zoo1-1 Starting Stderr: Container roottestclusterdiscoverypassword-gw7-zoo3-1 Starting Stderr: Container roottestclusterdiscoverypassword-gw7-zoo1-1 Started Stderr: Container roottestclusterdiscoverypassword-gw7-zoo2-1 Started Stderr: Container roottestclusterdiscoverypassword-gw7-zoo3-1 Started Stderr:time="2025-04-02T03:34:46Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T03:34:46Z" level=debug msg="otel error" error="" Wait ZooKeeper to start get_instance_ip instance_name=zoo1 http://localhost:None "GET /v1.46/containers/roottestclusterdiscoverypassword-gw7-zoo1-1/json HTTP/1.1" 200 None get_kazoo_client: zoo1, ip:172.16.7.2, port:2181, use_ssl:False Connecting to 172.16.7.2(172.16.7.2):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Connecting to 172.16.1.3(172.16.1.3):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Command:[docker compose --env-file /ClickHouse/tests/integration/test_compatibility_merge_tree_settings/_instances-0-gw3/.env --project-name roottestcompatibilitymergetreesettings-gw3 --file /ClickHouse/tests/integration/test_compatibility_merge_tree_settings/_instances-0-gw3/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_compatibility_merge_tree_settings/_instances-0-gw3/node2/docker-compose.yml --file /ClickHouse/tests/integration/test_compatibility_merge_tree_settings/_instances-0-gw3/node3/docker-compose.yml down --volumes] Skipping toDecimal64OrNull Checking toDecimal32OrNull('foo') Executing query select toDecimal32OrNull('foo') on backward Executing query ALTER TABLE tbl MODIFY COLUMN x String on node1 Executing query INSERT INTO tbl SELECT number FROM numbers(500) on node Skipping toDecimal32OrNull Checking toDateTimeOrNull('foo') Executing query select toDateTimeOrNull('foo') on backward Connecting to 172.16.7.2(172.16.7.2):2181, use_ssl: False Connection dropped: socket connection error: No route to host Connection dropped: socket connection error: Connection refused Connection dropped: socket connection error: No route to host Executing query SELECT name, type FROM system.columns WHERE database='default' AND table='tbl' on node1 Stderr: Container roottestasyncconnecttomultipleips-gw8-node1-1 Stopping Stderr: Container roottestasyncconnecttomultipleips-gw8-node2-1 Stopping Stderr: Container roottestasyncconnecttomultipleips-gw8-node1-1 Stopped Stderr: Container roottestasyncconnecttomultipleips-gw8-node2-1 Stopped Stderr: Container roottestasyncconnecttomultipleips-gw8-zoo1-1 Stopping Stderr: Container roottestasyncconnecttomultipleips-gw8-zoo2-1 Stopping Stderr: Container roottestasyncconnecttomultipleips-gw8-zoo3-1 Stopping Stderr: Container roottestasyncconnecttomultipleips-gw8-zoo1-1 Stopped Stderr: Container roottestasyncconnecttomultipleips-gw8-zoo2-1 Stopped Stderr: Container roottestasyncconnecttomultipleips-gw8-zoo3-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_async_connect_to_multiple_ips/_instances-0-gw8/node1/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_async_connect_to_multiple_ips/_instances-0-gw8/node1/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Executing query WITH quantile(0.05)(cnt) as p05, quantile(0.95)(cnt) as p95, p95 - p05 as inter_percentile_range SELECT sum(cnt) as total_requests, count() as data_points, inter_percentile_range FROM ( SELECT count() as cnt FROM remote('node{1,2}', numbers(10)) GROUP BY number ) on node1 Command:[bash -c [ -f /ClickHouse/tests/integration/test_async_connect_to_multiple_ips/_instances-0-gw8/node2/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_async_connect_to_multiple_ips/_instances-0-gw8/node2/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_async_connect_to_multiple_ips/_instances-0-gw8/.env --project-name roottestasyncconnecttomultipleips-gw8 --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_net.yml --file /ClickHouse/tests/integration/test_async_connect_to_multiple_ips/_instances-0-gw8/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_async_connect_to_multiple_ips/_instances-0-gw8/node2/docker-compose.yml down --volumes] Executing query select toDateTimeOrNull('foo') on upstream Executing query SELECT name, type FROM system.columns WHERE database='default' AND table='tbl' on node2 Executing query BACKUP TABLE tbl TO Disk('backups', '0d4abe713f324bc6b013778cacfec174') SETTINGS id='0d4abe713f324bc6b013778cacfec174' ASYNC on node Connecting to 172.16.7.2(172.16.7.2):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Executing query SELECT status FROM system.backups WHERE id='0d4abe713f324bc6b013778cacfec174' on node OK toDateTimeOrNull Checking toDate32OrNull('foo') Executing query select toDate32OrNull('foo') on backward Executing query BACKUP TABLE tbl ON CLUSTER 'cluster' TO Disk('backups', '21') on node2 Command:[docker compose --env-file /ClickHouse/tests/integration/test_backward_compatibility/_instances-cte_distributed-0-gw6/.env --project-name roottestbackwardcompatibilityctedistributed-gw6 --file /ClickHouse/tests/integration/test_backward_compatibility/_instances-cte_distributed-0-gw6/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_backward_compatibility/_instances-cte_distributed-0-gw6/node2/docker-compose.yml stop --timeout 20] [gw6] PASSED test_backward_compatibility/test_cte_distributed.py::test_cte_distributed Executing query SELECT count() FROM system.processes WHERE query_kind='Backup' AND query LIKE '%0d4abe713f324bc6b013778cacfec174%' on node Executing query select toDate32OrNull('foo') on upstream run container_id:roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] Command:[docker exec -u root roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 bash -c ps -C clickhouse] OK toDate32OrNull Checking toDateOrNull('foo') Executing query select toDateOrNull('foo') on backward Stdout: PID TTY TIME CMD Stdout: 10 ? 00:00:00 clickhouse run container_id:roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] Command:[docker exec -u root roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 bash -c pkill clickhouse] Executing query select toDateOrNull('foo') on upstream Executing query DROP TABLE tbl ON CLUSTER 'cluster' SYNC on node1 run container_id:roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Connecting to 172.16.7.2(172.16.7.2):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Stdout:10 OK toDateOrNull Checking toFloat64OrNull('foo') Executing query select toFloat64OrNull('foo') on backward Stderr: Container roottestcompatibilitymergetreesettings-gw3-node1-1 Stopping Stderr: Container roottestcompatibilitymergetreesettings-gw3-node2-1 Stopping Stderr: Container roottestcompatibilitymergetreesettings-gw3-node3-1 Stopping Stderr: Container roottestcompatibilitymergetreesettings-gw3-node1-1 Stopped Stderr: Container roottestcompatibilitymergetreesettings-gw3-node1-1 Removing Stderr: Container roottestcompatibilitymergetreesettings-gw3-node2-1 Stopped Stderr: Container roottestcompatibilitymergetreesettings-gw3-node2-1 Removing Stderr: Container roottestcompatibilitymergetreesettings-gw3-node3-1 Stopped Stderr: Container roottestcompatibilitymergetreesettings-gw3-node3-1 Removing Stderr: Container roottestcompatibilitymergetreesettings-gw3-node3-1 Removed Stderr: Container roottestcompatibilitymergetreesettings-gw3-node1-1 Removed Stderr: Container roottestcompatibilitymergetreesettings-gw3-node2-1 Removed Stderr: Container roottestcompatibilitymergetreesettings-gw3-zoo3-1 Stopping Stderr: Container roottestcompatibilitymergetreesettings-gw3-zoo1-1 Stopping Stderr: Container roottestcompatibilitymergetreesettings-gw3-zoo2-1 Stopping Stderr: Container roottestcompatibilitymergetreesettings-gw3-zoo3-1 Stopped Stderr: Container roottestcompatibilitymergetreesettings-gw3-zoo3-1 Removing Stderr: Container roottestcompatibilitymergetreesettings-gw3-zoo2-1 Stopped Stderr: Container roottestcompatibilitymergetreesettings-gw3-zoo2-1 Removing Stderr: Container roottestcompatibilitymergetreesettings-gw3-zoo1-1 Stopped Stderr: Container roottestcompatibilitymergetreesettings-gw3-zoo1-1 Removing Stderr: Container roottestcompatibilitymergetreesettings-gw3-zoo2-1 Removed Stderr: Container roottestcompatibilitymergetreesettings-gw3-zoo1-1 Removed Stderr: Container roottestcompatibilitymergetreesettings-gw3-zoo3-1 Removed Stderr: Network roottestcompatibilitymergetreesettings-gw3_default Removing Stderr: Network roottestcompatibilitymergetreesettings-gw3_default Removed Cleanup called Docker networks for project roottestcompatibilitymergetreesettings-gw3 are NETWORK ID NAME DRIVER SCOPE Executing query select toFloat64OrNull('foo') on upstream Docker containers for project roottestcompatibilitymergetreesettings-gw3 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestcompatibilitymergetreesettings-gw3 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestcompatibilitymergetreesettings-gw3-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestcompatibilitymergetreesettings-gw3 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Executing query RESTORE TABLE tbl ON CLUSTER 'cluster' FROM Disk('backups', '21') SETTINGS replica_num_in_backup=1 on node1 Connecting to 172.16.2.3(172.16.2.3):2181, use_ssl: False Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:1 Volumes pruned: 1 test_build_sets_from_multiple_threads/test.py::test_set Running tests in /ClickHouse/tests/integration/test_build_sets_from_multiple_threads/test.py Cluster start called. is_up=False Docker networks for project roottestbuildsetsfrommultiplethreads-gw3 are NETWORK ID NAME DRIVER SCOPE OK toFloat64OrNull Checking toInt256OrNull('foo') Executing query select toInt256OrNull('foo') on backward Connecting to 172.16.2.3(172.16.2.3):2181, use_ssl: False Docker containers for project roottestbuildsetsfrommultiplethreads-gw3 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestbuildsetsfrommultiplethreads-gw3 are DRIVER VOLUME NAME Cleanup called Docker networks for project roottestbuildsetsfrommultiplethreads-gw3 are NETWORK ID NAME DRIVER SCOPE Executing query select toInt256OrNull('foo') on upstream Docker containers for project roottestbuildsetsfrommultiplethreads-gw3 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestbuildsetsfrommultiplethreads-gw3 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestbuildsetsfrommultiplethreads-gw3-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestbuildsetsfrommultiplethreads-gw3 Trying to prune unused networks... Executing query SELECT name, type FROM system.columns WHERE database='default' AND table='tbl' on node1 Trying to prune unused images... Command:[docker image prune -f] OK toInt256OrNull Checking toInt128OrNull('foo') Executing query select toInt128OrNull('foo') on backward Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:1 Volumes pruned: 1 Setup directory for instance: node Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files [] to /ClickHouse/tests/integration/test_build_sets_from_multiple_threads/_instances-0-gw3/node/configs/config.d Setup database dir /ClickHouse/tests/integration/test_build_sets_from_multiple_threads/_instances-0-gw3/node/database Setup logs dir /ClickHouse/tests/integration/test_build_sets_from_multiple_threads/_instances-0-gw3/node/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw'} stored in /ClickHouse/tests/integration/test_build_sets_from_multiple_threads/_instances-0-gw3/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Executing query SELECT name, type FROM system.columns WHERE database='default' AND table='tbl' on node2 http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_build_sets_from_multiple_threads/_instances-0-gw3/.env --project-name roottestbuildsetsfrommultiplethreads-gw3 --file /ClickHouse/tests/integration/test_build_sets_from_multiple_threads/_instances-0-gw3/node/docker-compose.yml pull] Executing query select toInt128OrNull('foo') on upstream Executing query DROP TABLE tbl ON CLUSTER 'cluster' SYNC on node1 Connecting to 172.16.1.3(172.16.1.3):2181, use_ssl: False Connection dropped: socket connection error: Connection refused OK toInt128OrNull Checking toInt64OrNull('foo') Executing query select toInt64OrNull('foo') on backward Executing query select toInt64OrNull('foo') on upstream Executing query RESTORE TABLE tbl ON CLUSTER 'cluster' FROM Disk('backups', '21') SETTINGS replica_num_in_backup=2 on node2 OK toInt64OrNull Checking toUInt128OrNull('foo') Executing query select toUInt128OrNull('foo') on backward Executing query select toUInt128OrNull('foo') on upstream Executing query SELECT name, type FROM system.columns WHERE database='default' AND table='tbl' on node1 OK toUInt128OrNull Checking toUInt64OrNull('foo') Executing query select toUInt64OrNull('foo') on backward Connecting to 172.16.7.2(172.16.7.2):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Stderr: Container roottestasyncconnecttomultipleips-gw8-node1-1 Stopping Stderr: Container roottestasyncconnecttomultipleips-gw8-node2-1 Stopping Stderr: Container roottestasyncconnecttomultipleips-gw8-node1-1 Stopped Stderr: Container roottestasyncconnecttomultipleips-gw8-node1-1 Removing Stderr: Container roottestasyncconnecttomultipleips-gw8-node2-1 Stopped Stderr: Container roottestasyncconnecttomultipleips-gw8-node2-1 Removing Stderr: Container roottestasyncconnecttomultipleips-gw8-node1-1 Removed Stderr: Container roottestasyncconnecttomultipleips-gw8-node2-1 Removed Stderr: Container roottestasyncconnecttomultipleips-gw8-zoo1-1 Stopping Stderr: Container roottestasyncconnecttomultipleips-gw8-zoo3-1 Stopping Stderr: Container roottestasyncconnecttomultipleips-gw8-zoo2-1 Stopping Stderr: Container roottestasyncconnecttomultipleips-gw8-zoo1-1 Stopped Stderr: Container roottestasyncconnecttomultipleips-gw8-zoo1-1 Removing Executing query SELECT name, type FROM system.columns WHERE database='default' AND table='tbl' on node2 Stderr: Container roottestasyncconnecttomultipleips-gw8-zoo3-1 Stopped Stderr: Container roottestasyncconnecttomultipleips-gw8-zoo3-1 Removing Stderr: Container roottestasyncconnecttomultipleips-gw8-zoo2-1 Stopped Stderr: Container roottestasyncconnecttomultipleips-gw8-zoo2-1 Removing Stderr: Container roottestasyncconnecttomultipleips-gw8-zoo2-1 Removed Stderr: Container roottestasyncconnecttomultipleips-gw8-zoo3-1 Removed Stderr: Container roottestasyncconnecttomultipleips-gw8-zoo1-1 Removed Stderr: Network roottestasyncconnecttomultipleips-gw8_default Removing Stderr: Network roottestasyncconnecttomultipleips-gw8_default Removed Cleanup called Executing query select toUInt64OrNull('foo') on upstream Docker networks for project roottestasyncconnecttomultipleips-gw8 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestasyncconnecttomultipleips-gw8 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestasyncconnecttomultipleips-gw8 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestasyncconnecttomultipleips-gw8-.*-1$' --format '{{.ID}}:{{.Names}}'] Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw0] PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_def OK toUInt64OrNull Checking toUInt8OrNull('foo') Executing query select toUInt8OrNull('foo') on backward Unstopped containers: {} No running containers for project: roottestasyncconnecttomultipleips-gw8 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:1 Volumes pruned: 1 test_broken_part_during_merge/test.py::test_merge_and_part_corruption Running tests in /ClickHouse/tests/integration/test_broken_part_during_merge/test.py Cluster start called. is_up=False Executing query select toUInt8OrNull('foo') on upstream Docker networks for project roottestbrokenpartduringmerge-gw8 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestbrokenpartduringmerge-gw8 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestbrokenpartduringmerge-gw8 are DRIVER VOLUME NAME Cleanup called Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Docker networks for project roottestbrokenpartduringmerge-gw8 are NETWORK ID NAME DRIVER SCOPE OK toUInt8OrNull Checking toIPv6OrZero('foo') Executing query select toIPv6OrZero('foo') on backward Docker containers for project roottestbrokenpartduringmerge-gw8 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestbrokenpartduringmerge-gw8 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestbrokenpartduringmerge-gw8-.*-1$' --format '{{.ID}}:{{.Names}}'] run container_id:roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Unstopped containers: {} No running containers for project: roottestbrokenpartduringmerge-gw8 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Executing query select toIPv6OrZero('foo') on upstream Stdout:10 Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:1 Volumes pruned: 1 Setup directory for instance: node1 Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files [] to /ClickHouse/tests/integration/test_broken_part_during_merge/_instances-0-gw8/node1/configs/config.d Setup database dir /ClickHouse/tests/integration/test_broken_part_during_merge/_instances-0-gw8/node1/database Setup logs dir /ClickHouse/tests/integration/test_broken_part_during_merge/_instances-0-gw8/node1/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'keeper_binary': '/clickhouse', 'keeper_cmd_prefix': 'clickhouse keeper', 'image': 'altinityinfra/integration-test:8b2301119731', 'user': '0', 'keeper_fs': 'bind', 'keeper_logs_dir1': '/ClickHouse/tests/integration/test_broken_part_during_merge/_instances-0-gw8/keeper1/log', 'keeper_config_dir1': '/ClickHouse/tests/integration/test_broken_part_during_merge/_instances-0-gw8/keeper1/config', 'keeper_db_dir1': '/ClickHouse/tests/integration/test_broken_part_during_merge/_instances-0-gw8/keeper1/coordination', 'keeper_logs_dir2': '/ClickHouse/tests/integration/test_broken_part_during_merge/_instances-0-gw8/keeper2/log', 'keeper_config_dir2': '/ClickHouse/tests/integration/test_broken_part_during_merge/_instances-0-gw8/keeper2/config', 'keeper_db_dir2': '/ClickHouse/tests/integration/test_broken_part_during_merge/_instances-0-gw8/keeper2/coordination', 'keeper_logs_dir3': '/ClickHouse/tests/integration/test_broken_part_during_merge/_instances-0-gw8/keeper3/log', 'keeper_config_dir3': '/ClickHouse/tests/integration/test_broken_part_during_merge/_instances-0-gw8/keeper3/config', 'keeper_db_dir3': '/ClickHouse/tests/integration/test_broken_part_during_merge/_instances-0-gw8/keeper3/coordination'} stored in /ClickHouse/tests/integration/test_broken_part_during_merge/_instances-0-gw8/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_broken_part_during_merge/_instances-0-gw8/.env --project-name roottestbrokenpartduringmerge-gw8 --file /ClickHouse/tests/integration/test_broken_part_during_merge/_instances-0-gw8/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml pull] OK toIPv6OrZero Checking toIPv4OrZero('foo') Executing query select toIPv4OrZero('foo') on backward Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Executing query select toIPv4OrZero('foo') on upstream OK toIPv4OrZero Checking toDecimal256OrZero('foo') Executing query select toDecimal256OrZero('foo') on backward Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Skipping toDecimal256OrZero Checking toDecimal64OrZero('foo') Executing query select toDecimal64OrZero('foo') on backward Skipping toDecimal64OrZero Checking toDecimal32OrZero('foo') Executing query select toDecimal32OrZero('foo') on backward Skipping toDecimal32OrZero Checking toDateOrZero('foo') Executing query select toDateOrZero('foo') on backward Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Executing query select toDateOrZero('foo') on upstream Command:[docker compose --env-file /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw1/.env --project-name roottestclusterallreplicas-gw1 --file /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw1/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw1/node2/docker-compose.yml stop --timeout 20] [gw1] PASSED test_cluster_all_replicas/test.py::test_skip_unavailable_replica[two_shards_three_nodes] OK toDateOrZero Checking toFloat64OrZero('foo') Executing query select toFloat64OrZero('foo') on backward test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_insert Executing query CREATE TABLE tbl ON CLUSTER 'cluster' (x UInt32) ENGINE=ReplicatedMergeTree('/clickhouse/tables/tbl/', '{replica}')ORDER BY x on node1 Executing query select toFloat64OrZero('foo') on upstream Executing query INSERT INTO tbl VALUES (111) on node1 OK toFloat64OrZero Checking toFloat32OrZero('foo') Executing query select toFloat32OrZero('foo') on backward Executing query INSERT INTO tbl VALUES (222) on node2 Executing query select toFloat32OrZero('foo') on upstream Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster' tbl on node1 OK toFloat32OrZero Checking toInt128OrZero('foo') Executing query select toInt128OrZero('foo') on backward Executing query select toInt128OrZero('foo') on upstream Executing query SYSTEM STOP REPLICATED SENDS ON CLUSTER 'cluster' tbl on node1 OK toInt128OrZero Checking toInt64OrZero('foo') Executing query select toInt64OrZero('foo') on backward Executing query select toInt64OrZero('foo') on upstream Executing query INSERT INTO tbl VALUES (333) on node1 Executing query INSERT INTO tbl VALUES (444) on node2 OK toInt64OrZero Checking toInt16OrZero('foo') Executing query select toInt16OrZero('foo') on backward run container_id:roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query BACKUP TABLE tbl ON CLUSTER 'cluster' TO Disk('backups', '22') on node1 Executing query select toInt16OrZero('foo') on upstream Stdout:10 OK toInt16OrZero Checking toUInt128OrZero('foo') Executing query select toUInt128OrZero('foo') on backward Connecting to 172.16.7.2(172.16.7.2):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost Executing query select toUInt128OrZero('foo') on upstream Executing query DROP TABLE tbl ON CLUSTER 'cluster' SYNC on node1 Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED get_instance_ip instance_name=zoo2 OK toUInt128OrZero Checking toUInt32OrZero('foo') Executing query select toUInt32OrZero('foo') on backward http://localhost:None "GET /v1.46/containers/roottestclusterdiscoverypassword-gw7-zoo2-1/json HTTP/1.1" 200 None get_kazoo_client: zoo2, ip:172.16.7.3, port:2181, use_ssl:False Connecting to 172.16.7.3(172.16.7.3):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost Executing query RESTORE TABLE tbl ON CLUSTER 'cluster' FROM Disk('backups', '22') on node1 Executing query select toUInt32OrZero('foo') on upstream Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED get_instance_ip instance_name=zoo3 http://localhost:None "GET /v1.46/containers/roottestclusterdiscoverypassword-gw7-zoo3-1/json HTTP/1.1" 200 None get_kazoo_client: zoo3, ip:172.16.7.4, port:2181, use_ssl:False Connecting to 172.16.7.4(172.16.7.4):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost OK toUInt32OrZero Checking toString('foo') Executing query select toString('foo') on backward Executing query select toString('foo') on upstream Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED All instances of ZooKeeper started: ('zoo1', 'zoo2', 'zoo3') ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_cluster_discovery/_instances-password-0-gw7/.env --project-name roottestclusterdiscoverypassword-gw7 --file /ClickHouse/tests/integration/test_cluster_discovery/_instances-password-0-gw7/node0/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_cluster_discovery/_instances-password-0-gw7/node1/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_cluster_discovery/_instances-password-0-gw7/.env --project-name roottestclusterdiscoverypassword-gw7 --file /ClickHouse/tests/integration/test_cluster_discovery/_instances-password-0-gw7/node0/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_cluster_discovery/_instances-password-0-gw7/node1/docker-compose.yml up -d --no-recreate] Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster' tbl on node1 OK toString Checking toDateTime('foo') Executing query select toDateTime('foo') on backward Skipping toDateTime Checking toDecimal128('foo') Executing query select toDecimal128('foo') on backward Connecting to 172.16.1.3(172.16.1.3):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost Skipping toDecimal128 Checking toDecimal64('foo') Executing query select toDecimal64('foo') on backward Executing query SELECT * FROM tbl ORDER BY x on node1 Skipping toDecimal64 Checking toFloat64('foo') Executing query select toFloat64('foo') on backward Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED get_instance_ip instance_name=zoo2 http://localhost:None "GET /v1.46/containers/roottestbackwardcompatibilityconvertordinary-gw5-zoo2-1/json HTTP/1.1" 200 None get_kazoo_client: zoo2, ip:172.16.1.4, port:2181, use_ssl:False Connecting to 172.16.1.4(172.16.1.4):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost Executing query SELECT * FROM tbl ORDER BY x on node2 Skipping toFloat64 Checking toFloat32('foo') Executing query select toFloat32('foo') on backward Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED get_instance_ip instance_name=zoo3 http://localhost:None "GET /v1.46/containers/roottestbackwardcompatibilityconvertordinary-gw5-zoo3-1/json HTTP/1.1" 200 None get_kazoo_client: zoo3, ip:172.16.1.2, port:2181, use_ssl:False Connecting to 172.16.1.2(172.16.1.2):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost Skipping toFloat32 Checking toInt128('foo') Executing query select toInt128('foo') on backward Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED All instances of ZooKeeper started: ('zoo1', 'zoo2', 'zoo3') ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_backward_compatibility/_instances-convert_ordinary-0-gw5/.env --project-name roottestbackwardcompatibilityconvertordinary-gw5 --file /ClickHouse/tests/integration/test_backward_compatibility/_instances-convert_ordinary-0-gw5/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_backward_compatibility/_instances-convert_ordinary-0-gw5/.env --project-name roottestbackwardcompatibilityconvertordinary-gw5 --file /ClickHouse/tests/integration/test_backward_compatibility/_instances-convert_ordinary-0-gw5/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml up -d --no-recreate] Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw0] PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_insert Skipping toInt128 Stderr: Container roottestclusterdiscoverypassword-gw7-zoo2-1 Running Checking toInt32('foo') Stderr: Container roottestclusterdiscoverypassword-gw7-zoo1-1 Running Stderr: Container roottestclusterdiscoverypassword-gw7-zoo3-1 Running Stderr: Container roottestclusterdiscoverypassword-gw7-node0-1 Creating Stderr: Container roottestclusterdiscoverypassword-gw7-node1-1 Creating Stderr: Container roottestclusterdiscoverypassword-gw7-node0-1 Created Stderr: Container roottestclusterdiscoverypassword-gw7-node1-1 Created Stderr: Container roottestclusterdiscoverypassword-gw7-node0-1 Starting Stderr: Container roottestclusterdiscoverypassword-gw7-node1-1 Starting Stderr: Container roottestclusterdiscoverypassword-gw7-node1-1 Started Stderr: Container roottestclusterdiscoverypassword-gw7-node0-1 Started ClickHouse instance created get_instance_ip instance_name=node0 Executing query select toInt32('foo') on backward http://localhost:None "GET /v1.46/containers/roottestclusterdiscoverypassword-gw7-node0-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node0 http://localhost:None "GET /v1.46/containers/roottestclusterdiscoverypassword-gw7-node0-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node0, ip: 172.16.7.6... http://localhost:None "GET /v1.46/containers/roottestclusterdiscoverypassword-gw7-node0-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/3ffd4a3ffbc3b2d1a4e634b36771191f58bdd3d5f8ddc1eb44da89fd4614a8a7/json HTTP/1.1" 200 None Skipping toInt32 Checking toInt8('foo') Executing query select toInt8('foo') on backward http://localhost:None "GET /v1.46/containers/3ffd4a3ffbc3b2d1a4e634b36771191f58bdd3d5f8ddc1eb44da89fd4614a8a7/json HTTP/1.1" 200 None Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Skipping toInt8 Checking toUInt256('foo') Executing query select toUInt256('foo') on backward run container_id:roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:10 Skipping toUInt256 Checking toUInt64('foo') Executing query select toUInt64('foo') on backward http://localhost:None "GET /v1.46/containers/3ffd4a3ffbc3b2d1a4e634b36771191f58bdd3d5f8ddc1eb44da89fd4614a8a7/json HTTP/1.1" 200 None Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Stderr: Container roottestbackwardcompatibilityconvertordinary-gw5-zoo3-1 Running Stderr: Container roottestbackwardcompatibilityconvertordinary-gw5-zoo2-1 Running Stderr: Container roottestbackwardcompatibilityconvertordinary-gw5-zoo1-1 Running Stderr: Container roottestbackwardcompatibilityconvertordinary-gw5-node-1 Creating Stderr: Container roottestbackwardcompatibilityconvertordinary-gw5-node-1 Created Stderr: Container roottestbackwardcompatibilityconvertordinary-gw5-node-1 Starting Stderr: Container roottestbackwardcompatibilityconvertordinary-gw5-node-1 Started ClickHouse instance created get_instance_ip instance_name=node http://localhost:None "GET /v1.46/containers/roottestbackwardcompatibilityconvertordinary-gw5-node-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node http://localhost:None "GET /v1.46/containers/roottestbackwardcompatibilityconvertordinary-gw5-node-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node, ip: 172.16.1.5... Skipping toUInt64 Checking toDecimal32('foo') http://localhost:None "GET /v1.46/containers/roottestbackwardcompatibilityconvertordinary-gw5-node-1/json HTTP/1.1" 200 None Executing query select toDecimal32('foo') on backward http://localhost:None "GET /v1.46/containers/3ffd4a3ffbc3b2d1a4e634b36771191f58bdd3d5f8ddc1eb44da89fd4614a8a7/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/fd8ccd2faccb217404faf75f87488020e20c758d1379a20fcf33a41d2cf21873/json HTTP/1.1" 200 None Skipping toDecimal32 Checking toUInt32('foo') Executing query select toUInt32('foo') on backward Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 http://localhost:None "GET /v1.46/containers/3ffd4a3ffbc3b2d1a4e634b36771191f58bdd3d5f8ddc1eb44da89fd4614a8a7/json HTTP/1.1" 200 None ClickHouse node0 started http://localhost:None "GET /v1.46/containers/fd8ccd2faccb217404faf75f87488020e20c758d1379a20fcf33a41d2cf21873/json HTTP/1.1" 200 None get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottestclusterdiscoverypassword-gw7-node1-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottestclusterdiscoverypassword-gw7-node1-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node1, ip: 172.16.7.5... http://localhost:None "GET /v1.46/containers/roottestclusterdiscoverypassword-gw7-node1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/459c9aaded47fbd63e24e3f86396b8e0487c03031ddd540d19818d273f286010/json HTTP/1.1" 200 None ClickHouse node1 started Executing query SELECT count() FROM system.clusters WHERE cluster = 'test_auto_cluster_with_pwd' on node0 Skipping toUInt32 Checking toUInt8('foo') Executing query select toUInt8('foo') on backward http://localhost:None "GET /v1.46/containers/fd8ccd2faccb217404faf75f87488020e20c758d1379a20fcf33a41d2cf21873/json HTTP/1.1" 200 None Skipping toUInt8 Checking blockSize('foo') Executing query select blockSize('foo') on backward Stderr: Container roottestbackwardcompatibilityctedistributed-gw6-node2-1 Stopping Stderr: Container roottestbackwardcompatibilityctedistributed-gw6-node1-1 Stopping Stderr: Container roottestbackwardcompatibilityctedistributed-gw6-node2-1 Stopped Stderr: Container roottestbackwardcompatibilityctedistributed-gw6-node1-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_backward_compatibility/_instances-cte_distributed-0-gw6/node1/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_backward_compatibility/_instances-cte_distributed-0-gw6/node1/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_backward_compatibility/_instances-cte_distributed-0-gw6/node2/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_backward_compatibility/_instances-cte_distributed-0-gw6/node2/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_backward_compatibility/_instances-cte_distributed-0-gw6/.env --project-name roottestbackwardcompatibilityctedistributed-gw6 --file /ClickHouse/tests/integration/test_backward_compatibility/_instances-cte_distributed-0-gw6/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_backward_compatibility/_instances-cte_distributed-0-gw6/node2/docker-compose.yml down --volumes] Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 http://localhost:None "GET /v1.46/containers/fd8ccd2faccb217404faf75f87488020e20c758d1379a20fcf33a41d2cf21873/json HTTP/1.1" 200 None Skipping blockSize Checking toUInt16('foo') Executing query select toUInt16('foo') on backward Executing query SELECT count() FROM system.clusters WHERE cluster = 'test_auto_cluster_with_pwd' on node1 Skipping toUInt16 Checking isNotNull('foo') Executing query select isNotNull('foo') on backward http://localhost:None "GET /v1.46/containers/fd8ccd2faccb217404faf75f87488020e20c758d1379a20fcf33a41d2cf21873/json HTTP/1.1" 200 None ClickHouse node started Executing query CREATE DATABASE default2 ENGINE=Ordinary on node Executing query SELECT sum(number) FROM clusterAllReplicas('test_auto_cluster_with_pwd', numbers(3)) GROUP BY hostname() on node0 test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_merge Executing query CREATE TABLE tbl ON CLUSTER 'cluster' (x UInt32) ENGINE=ReplicatedMergeTree('/clickhouse/tables/tbl/', '{replica}')ORDER BY x on node1 Executing query CREATE TABLE default2.t(date Date, id UInt32) ENGINE = MergeTree PARTITION BY toYYYYMM(date) ORDER BY id on node Executing query select isNotNull('foo') on upstream Executing query SELECT sum(number) FROM clusterAllReplicas('test_auto_cluster_with_wrong_pwd', numbers(3)) GROUP BY hostname() on node0 Executing query INSERT INTO default2.t VALUES (today(), 1) on node OK isNotNull Checking UUIDStringToNum('foo') Executing query select UUIDStringToNum('foo') on backward Executing query SELECT sum(number) FROM clusterAllReplicas('test_auto_cluster_with_secret', numbers(3)) GROUP BY hostname() on node0 Executing query SYSTEM STOP MERGES ON CLUSTER 'cluster' tbl on node1 Executing query INSERT INTO default2.t SELECT number % 1000, number FROM system.numbers LIMIT 1000000 on node Executing query select UUIDStringToNum('foo') on upstream Executing query SELECT sum(number) FROM clusterAllReplicas('test_auto_cluster_with_wrong_secret', numbers(3)) GROUP BY hostname() on node0 OK UUIDStringToNum Checking roundDown('foo') Executing query select roundDown('foo') on backward run container_id:roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query INSERT INTO tbl VALUES (111) on node1 Skipping roundDown Checking trunc('foo') Executing query select trunc('foo') on backward Command:[docker compose --env-file /ClickHouse/tests/integration/test_cluster_discovery/_instances-password-0-gw7/.env --project-name roottestclusterdiscoverypassword-gw7 --file /ClickHouse/tests/integration/test_cluster_discovery/_instances-password-0-gw7/node0/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_cluster_discovery/_instances-password-0-gw7/node1/docker-compose.yml stop --timeout 20] [gw7] PASSED test_cluster_discovery/test_password.py::test_connect_with_password Stderr: Container roottestbackwardcompatibilityctedistributed-gw6-node1-1 Stopping Stderr: Container roottestbackwardcompatibilityctedistributed-gw6-node2-1 Stopping Stderr: Container roottestbackwardcompatibilityctedistributed-gw6-node1-1 Stopped Stderr: Container roottestbackwardcompatibilityctedistributed-gw6-node1-1 Removing Stderr: Container roottestbackwardcompatibilityctedistributed-gw6-node2-1 Stopped Stderr: Container roottestbackwardcompatibilityctedistributed-gw6-node2-1 Removing Stderr: Container roottestbackwardcompatibilityctedistributed-gw6-node1-1 Removed Stdout:10 Stderr: Container roottestbackwardcompatibilityctedistributed-gw6-node2-1 Removed Stderr: Network roottestbackwardcompatibilityctedistributed-gw6_default Removing Stderr: Network roottestbackwardcompatibilityctedistributed-gw6_default Removed Cleanup called Executing query SELECT count() FROM default2.t on node Docker networks for project roottestbackwardcompatibilityctedistributed-gw6 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestbackwardcompatibilityctedistributed-gw6 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Executing query INSERT INTO tbl VALUES (222) on node1 Skipping trunc Checking ceil('foo') Executing query select ceil('foo') on backward Docker volumes for project roottestbackwardcompatibilityctedistributed-gw6 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestbackwardcompatibilityctedistributed-gw6-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestbackwardcompatibilityctedistributed-gw6 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Executing query SYSTEM SYNC REPLICA tbl on node2 Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Skipping ceil Checking roundBankers('foo') Executing query select roundBankers('foo') on backward Executing query SELECT sum(id) FROM default2.t on node Stdout:1 Volumes pruned: 1 test_cleanup_after_start/test.py::test_old_dirs_cleanup Running tests in /ClickHouse/tests/integration/test_cleanup_after_start/test.py Cluster start called. is_up=False Docker networks for project roottestcleanupafterstart-gw6 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestcleanupafterstart-gw6 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Executing query SYSTEM START MERGES tbl on node2 Skipping roundBankers Docker volumes for project roottestcleanupafterstart-gw6 are DRIVER VOLUME NAME Cleanup called Checking round('foo') Executing query select round('foo') on backward Executing query SELECT date, count(), sum(id) FROM default2.t GROUP BY date ORDER BY date LIMIT 2 on node Docker networks for project roottestcleanupafterstart-gw6 are NETWORK ID NAME DRIVER SCOPE Skipping round Checking regionToName('foo') Docker containers for project roottestcleanupafterstart-gw6 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Executing query select regionToName('foo') on backward Docker volumes for project roottestcleanupafterstart-gw6 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestcleanupafterstart-gw6-.*-1$' --format '{{.ID}}:{{.Names}}'] Executing query OPTIMIZE TABLE tbl FINAL on node2 Unstopped containers: {} No running containers for project: roottestcleanupafterstart-gw6 Trying to prune unused networks... Executing query SYSTEM FLUSH LOGS on node Trying to prune unused images... Command:[docker image prune -f] Skipping regionToName Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Checking regionIn('foo') Command:[docker volume ls | wc -l] Executing query select regionIn('foo') on backward Stdout:1 Volumes pruned: 1 Setup directory for instance: node1 Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files [] to /ClickHouse/tests/integration/test_cleanup_after_start/_instances-0-gw6/node1/configs/config.d Setup database dir /ClickHouse/tests/integration/test_cleanup_after_start/_instances-0-gw6/node1/database Setup logs dir /ClickHouse/tests/integration/test_cleanup_after_start/_instances-0-gw6/node1/logs Entrypoint cmd: bash -c "trap 'pkill tail' INT TERM; clickhouse server --config-file=/etc/clickhouse-server/config.xml --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log --daemon -- ; coproc tail -f /dev/null; wait $$!" Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'keeper_binary': '/clickhouse', 'keeper_cmd_prefix': 'clickhouse keeper', 'image': 'altinityinfra/integration-test:8b2301119731', 'user': '0', 'keeper_fs': 'bind', 'keeper_logs_dir1': '/ClickHouse/tests/integration/test_cleanup_after_start/_instances-0-gw6/keeper1/log', 'keeper_config_dir1': '/ClickHouse/tests/integration/test_cleanup_after_start/_instances-0-gw6/keeper1/config', 'keeper_db_dir1': '/ClickHouse/tests/integration/test_cleanup_after_start/_instances-0-gw6/keeper1/coordination', 'keeper_logs_dir2': '/ClickHouse/tests/integration/test_cleanup_after_start/_instances-0-gw6/keeper2/log', 'keeper_config_dir2': '/ClickHouse/tests/integration/test_cleanup_after_start/_instances-0-gw6/keeper2/config', 'keeper_db_dir2': '/ClickHouse/tests/integration/test_cleanup_after_start/_instances-0-gw6/keeper2/coordination', 'keeper_logs_dir3': '/ClickHouse/tests/integration/test_cleanup_after_start/_instances-0-gw6/keeper3/log', 'keeper_config_dir3': '/ClickHouse/tests/integration/test_cleanup_after_start/_instances-0-gw6/keeper3/config', 'keeper_db_dir3': '/ClickHouse/tests/integration/test_cleanup_after_start/_instances-0-gw6/keeper3/coordination'} stored in /ClickHouse/tests/integration/test_cleanup_after_start/_instances-0-gw6/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_cleanup_after_start/_instances-0-gw6/.env --project-name roottestcleanupafterstart-gw6 --file /ClickHouse/tests/integration/test_cleanup_after_start/_instances-0-gw6/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml pull] Skipping regionIn Checking toUUID('foo') Executing query select toUUID('foo') on backward Executing query BACKUP TABLE tbl ON CLUSTER 'cluster' TO Disk('backups', '23') on node1 Skipping toUUID Checking regionToCountry('foo') Executing query select regionToCountry('foo') on backward Executing query SHOW TABLES FROM system on node Skipping regionToCountry Checking regionToArea('foo') Executing query select regionToArea('foo') on backward Executing query DROP TABLE tbl ON CLUSTER 'cluster' SYNC on node1 Executing query SHOW TABLES FROM system on node Skipping regionToArea Checking regionToCity('foo') Executing query select regionToCity('foo') on backward Executing query SYSTEM FLUSH LOGS on node Skipping regionToCity Checking cutQueryString('foo') Executing query select cutQueryString('foo') on backward Executing query RESTORE TABLE tbl ON CLUSTER 'cluster' FROM Disk('backups', '23') on node1 Executing query SELECT count() != 0 FROM system.query_log on node Executing query select cutQueryString('foo') on upstream Executing query SELECT count() != 0 FROM system.part_log on node OK cutQueryString Checking BLAKE3('foo') Executing query select BLAKE3('foo') on backward run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill -15 clickhouse'] Command:[docker exec -u root roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c pkill -15 clickhouse] Executing query select BLAKE3('foo') on upstream Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster' tbl on node1 run container_id:roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse server' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c ps ax | grep 'clickhouse server' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:10 Stdout:8 OK BLAKE3 Checking wyHash64('foo') Executing query select wyHash64('foo') on backward Executing query SELECT * FROM tbl ORDER BY x on node1 Executing query select wyHash64('foo') on upstream Stderr: Container roottestclusterdiscoverypassword-gw7-node0-1 Stopping Stderr: Container roottestclusterdiscoverypassword-gw7-node1-1 Stopping Stderr: Container roottestclusterdiscoverypassword-gw7-node0-1 Stopped Stderr: Container roottestclusterdiscoverypassword-gw7-node1-1 Stopped Stderr: Container roottestclusterdiscoverypassword-gw7-zoo3-1 Stopping Stderr: Container roottestclusterdiscoverypassword-gw7-zoo1-1 Stopping Stderr: Container roottestclusterdiscoverypassword-gw7-zoo2-1 Stopping Stderr: Container roottestclusterdiscoverypassword-gw7-zoo2-1 Stopped Stderr: Container roottestclusterdiscoverypassword-gw7-zoo3-1 Stopped Stderr: Container roottestclusterdiscoverypassword-gw7-zoo1-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_cluster_discovery/_instances-password-0-gw7/node0/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_cluster_discovery/_instances-password-0-gw7/node0/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_cluster_discovery/_instances-password-0-gw7/node1/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_cluster_discovery/_instances-password-0-gw7/node1/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Executing query SELECT * FROM tbl ORDER BY x on node2 Command:[docker compose --env-file /ClickHouse/tests/integration/test_cluster_discovery/_instances-password-0-gw7/.env --project-name roottestclusterdiscoverypassword-gw7 --file /ClickHouse/tests/integration/test_cluster_discovery/_instances-password-0-gw7/node0/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_cluster_discovery/_instances-password-0-gw7/node1/docker-compose.yml down --volumes] OK wyHash64 Checking xxHash32('foo') Executing query select xxHash32('foo') on backward Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw0] PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_merge run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse server' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c ps ax | grep 'clickhouse server' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select xxHash32('foo') on upstream Stdout:8 Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 OK xxHash32 Checking javaHashUTF16LE('foo') Executing query select javaHashUTF16LE('foo') on backward Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Skipping javaHashUTF16LE Checking javaHash('foo') Executing query select javaHash('foo') on backward Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select javaHash('foo') on upstream run container_id:roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] OK javaHash Checking URLHash('foo') Executing query select URLHash('foo') on backward Stderr: Container roottestclusterdiscoverypassword-gw7-node1-1 Stopping Stderr: Container roottestclusterdiscoverypassword-gw7-node0-1 Stopping Stderr: Container roottestclusterdiscoverypassword-gw7-node1-1 Stopped Stderr: Container roottestclusterdiscoverypassword-gw7-node1-1 Removing Stderr: Container roottestclusterdiscoverypassword-gw7-node0-1 Stopped Stderr: Container roottestclusterdiscoverypassword-gw7-node0-1 Removing Stderr: Container roottestclusterdiscoverypassword-gw7-node0-1 Removed Stderr: Container roottestclusterdiscoverypassword-gw7-node1-1 Removed Stderr: Container roottestclusterdiscoverypassword-gw7-zoo3-1 Stopping Stderr: Container roottestclusterdiscoverypassword-gw7-zoo1-1 Stopping Stderr: Container roottestclusterdiscoverypassword-gw7-zoo2-1 Stopping Stderr: Container roottestclusterdiscoverypassword-gw7-zoo3-1 Stopped Stderr: Container roottestclusterdiscoverypassword-gw7-zoo3-1 Removing Stderr: Container roottestclusterdiscoverypassword-gw7-zoo1-1 Stopped Stderr: Container roottestclusterdiscoverypassword-gw7-zoo1-1 Removing Stderr: Container roottestclusterdiscoverypassword-gw7-zoo2-1 Stopped Stderr: Container roottestclusterdiscoverypassword-gw7-zoo2-1 Removing Stderr: Container roottestclusterdiscoverypassword-gw7-zoo2-1 Removed Stderr: Container roottestclusterdiscoverypassword-gw7-zoo1-1 Removed Stderr: Container roottestclusterdiscoverypassword-gw7-zoo3-1 Removed Stderr: Network roottestclusterdiscoverypassword-gw7_default Removing Stderr: Network roottestclusterdiscoverypassword-gw7_default Removed Cleanup called Stdout:10 Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Docker networks for project roottestclusterdiscoverypassword-gw7 are NETWORK ID NAME DRIVER SCOPE run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse server' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c ps ax | grep 'clickhouse server' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select URLHash('foo') on upstream Docker containers for project roottestclusterdiscoverypassword-gw7 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestclusterdiscoverypassword-gw7 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestclusterdiscoverypassword-gw7-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestclusterdiscoverypassword-gw7 Trying to prune unused networks... Stdout:8 Trying to prune unused images... Command:[docker image prune -f] OK URLHash Checking metroHash64('foo') Executing query select metroHash64('foo') on backward Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:1 Volumes pruned: 1 Executing query select metroHash64('foo') on upstream test_backup_restore_on_cluster/test.py::test_replicated_table_with_uuid_in_zkpath Executing query CREATE TABLE tbl ON CLUSTER 'cluster' (x UInt8, y String) ENGINE=ReplicatedMergeTree('/clickhouse/tables/{uuid}','{replica}')ORDER BY x on node1 OK metroHash64 Checking farmFingerprint64('foo') Executing query select farmFingerprint64('foo') on backward Executing query INSERT INTO tbl VALUES (1, 'AA') on node1 Executing query select farmFingerprint64('foo') on upstream Executing query INSERT INTO tbl VALUES (2, 'BB') on node2 OK farmFingerprint64 Checking sipHash128ReferenceKeyed('foo') Executing query select sipHash128ReferenceKeyed('foo') on backward Executing query BACKUP TABLE tbl ON CLUSTER 'cluster' TO Disk('backups', '24') on node1 Skipping sipHash128ReferenceKeyed Checking sipHash128Reference('foo') Executing query select sipHash128Reference('foo') on backward Executing query select sipHash128Reference('foo') on upstream run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse server' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c ps ax | grep 'clickhouse server' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] OK sipHash128Reference Checking sipHash128('foo') Executing query select sipHash128('foo') on backward run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse server' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c ps ax | grep 'clickhouse server' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', 'cp /usr/bin/clickhouse /usr/share/clickhouse_original'] Command:[docker exec -u root roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c cp /usr/bin/clickhouse /usr/share/clickhouse_original] Executing query RESTORE TABLE tbl AS tbl2 ON CLUSTER 'cluster' FROM Disk('backups', '24') on node2 Executing query select sipHash128('foo') on upstream OK sipHash128 Checking toDate32('foo') Executing query select toDate32('foo') on backward Skipping toDate32 Checking lowCardinalityIndices('foo') Executing query select lowCardinalityIndices('foo') on backward Executing query INSERT INTO tbl2 VALUES (3, 'CC') on node1 Skipping lowCardinalityIndices Checking decrypt('foo') Executing query select decrypt('foo') on backward Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster' tbl on node1 Skipping decrypt Checking logTrace('foo') Executing query select logTrace('foo') on backward run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', 'cp /usr/share/clickhouse_fresh /usr/bin/clickhouse && chmod 777 /usr/bin/clickhouse'] Command:[docker exec -u root roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c cp /usr/share/clickhouse_fresh /usr/bin/clickhouse && chmod 777 /usr/bin/clickhouse] run container_id:roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select logTrace('foo') on upstream Stdout:10 Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster' tbl2 on node1 OK logTrace Checking toInt16OrNull('foo') Executing query select toInt16OrNull('foo') on backward Executing query select toInt16OrNull('foo') on upstream Executing query SELECT * FROM tbl ORDER BY x on node1 Executing query SELECT * FROM tbl2 ORDER BY x on node1 OK toInt16OrNull Checking timeSlots('foo') Executing query select timeSlots('foo') on backward Executing query SELECT * FROM tbl ORDER BY x on node2 run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', "echo 'restart_with_latest_version: From version' && /usr/share/clickhouse_original server --version && echo 'To version' /usr/share/clickhouse_fresh server --version"] Command:[docker exec roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c echo 'restart_with_latest_version: From version' && /usr/share/clickhouse_original server --version && echo 'To version' /usr/share/clickhouse_fresh server --version] Skipping timeSlots Checking stringToH3('foo') Executing query select stringToH3('foo') on backward Executing query SELECT * FROM tbl2 ORDER BY x on node2 Stdout:restart_with_latest_version: From version Stdout:ClickHouse server version 23.3.22.3 (official build). Stdout:To version /usr/share/clickhouse_fresh server --version run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', "if [ ! -f /var/lib/clickhouse/metadata/system.sql ]; then echo 'ATTACH DATABASE system ENGINE=Ordinary' > /var/lib/clickhouse/metadata/system.sql; fi"] Command:[docker exec roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c if [ ! -f /var/lib/clickhouse/metadata/system.sql ]; then echo 'ATTACH DATABASE system ENGINE=Ordinary' > /var/lib/clickhouse/metadata/system.sql; fi] Executing query select stringToH3('foo') on upstream run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', "if [ ! -f /var/lib/clickhouse/metadata/default.sql ]; then echo 'ATTACH DATABASE system ENGINE=Ordinary' > /var/lib/clickhouse/metadata/default.sql; fi"] Command:[docker exec roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c if [ ! -f /var/lib/clickhouse/metadata/default.sql ]; then echo 'ATTACH DATABASE system ENGINE=Ordinary' > /var/lib/clickhouse/metadata/default.sql; fi] Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw0] PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_with_uuid_in_zkpath run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', 'clickhouse server --config-file=/etc/clickhouse-server/config.xml --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log --daemon -- '] Command:[docker exec -u 0 roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c clickhouse server --config-file=/etc/clickhouse-server/config.xml --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log --daemon -- ] OK stringToH3 Checking asinh('foo') Executing query select asinh('foo') on backward run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Skipping asinh Checking cutWWW('foo') Executing query select cutWWW('foo') on backward Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Stdout:401 Executing query select 20 on node Executing query select cutWWW('foo') on upstream Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 OK cutWWW Checking toInt8OrZero('foo') Executing query select toInt8OrZero('foo') on backward Executing query select toInt8OrZero('foo') on upstream Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 OK toInt8OrZero Checking SHA1('foo') Executing query select SHA1('foo') on backward run container_id:roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select SHA1('foo') on upstream Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Stdout:10 OK SHA1 Checking MD5('foo') Executing query select MD5('foo') on backward Executing query select MD5('foo') on upstream test_backup_restore_on_cluster/test.py::test_required_privileges Executing query CREATE TABLE tbl ON CLUSTER 'cluster' (x UInt8) ENGINE=ReplicatedMergeTree('/clickhouse/tables/tbl/', '{replica}')ORDER BY x on node1 OK MD5 Checking halfMD5('foo') Executing query select halfMD5('foo') on backward Executing query select 20 on node Executing query select halfMD5('foo') on upstream Executing query SHOW CREATE DATABASE default2 on node Executing query INSERT INTO tbl VALUES (100) on node1 OK halfMD5 Checking equals('foo') Executing query select equals('foo') on backward Executing query SHOW CREATE DATABASE system on node Executing query CREATE USER u1 on node1 Skipping equals Checking geohashesInBox('foo') Executing query select geohashesInBox('foo') on backward Executing query SHOW TABLES FROM system on node Executing query GRANT CLUSTER ON *.* TO u1 on node1 Skipping geohashesInBox Checking JSONExtractKeys('foo') Executing query select JSONExtractKeys('foo') on backward Executing query SHOW TABLES FROM system on node Executing query BACKUP TABLE tbl ON CLUSTER 'cluster' TO Disk('backups', '25') on node1 Executing query select JSONExtractKeys('foo') on upstream Executing query SYSTEM FLUSH LOGS on node Executing query GRANT BACKUP ON tbl TO u1 on node1 OK JSONExtractKeys Checking JSONExtractKeysAndValuesRaw('foo') Executing query select JSONExtractKeysAndValuesRaw('foo') on backward Executing query BACKUP TABLE tbl ON CLUSTER 'cluster' TO Disk('backups', '25') on node1 Executing query select JSONExtractKeysAndValuesRaw('foo') on upstream OK JSONExtractKeysAndValuesRaw Checking toUInt16OrNull('foo') Executing query select toUInt16OrNull('foo') on backward Executing query SHOW TABLES FROM system on node Executing query select toUInt16OrNull('foo') on upstream Executing query SHOW TABLES FROM system on node Executing query DROP TABLE tbl ON CLUSTER 'cluster' SYNC on node1 OK toUInt16OrNull Checking JSONExtractRaw('foo') Executing query select JSONExtractRaw('foo') on backward Executing query SELECT count() != 0 FROM system.query_log_0 on node Executing query select JSONExtractRaw('foo') on upstream Executing query SELECT count() != 0 FROM system.part_log_0 on node Executing query RESTORE TABLE tbl AS tbl2 ON CLUSTER 'cluster' FROM Disk('backups', '25') on node1 OK JSONExtractRaw Checking JSONExtractString('foo') Executing query select JSONExtractString('foo') on backward Executing query SELECT date, count(), sum(id) FROM default2.t GROUP BY date ORDER BY date LIMIT 2 on node Executing query GRANT INSERT, CREATE TABLE ON tbl2 TO u1 on node1 Executing query select JSONExtractString('foo') on upstream run container_id:roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query SELECT name FROM system.databases ORDER BY name on node Executing query RESTORE TABLE tbl AS tbl2 ON CLUSTER 'cluster' FROM Disk('backups', '25') on node1 Stdout:10 OK JSONExtractString Checking SHA512('foo') Executing query select SHA512('foo') on backward run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', 'grep -a "" /var/log/clickhouse-server/clickhouse-server.log | wc -l'] Command:[docker exec roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c grep -a "" /var/log/clickhouse-server/clickhouse-server.log | wc -l] Stdout:0 run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', 'grep -a " Database" /var/log/clickhouse-server/clickhouse-server.log | wc -l'] Command:[docker exec roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c grep -a " Database" /var/log/clickhouse-server/clickhouse-server.log | wc -l] Executing query select SHA512('foo') on upstream Stdout:0 run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', 'grep -a "always include the lines below" /var/log/clickhouse-server/clickhouse-server.log | wc -l'] Command:[docker exec roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c grep -a "always include the lines below" /var/log/clickhouse-server/clickhouse-server.log | wc -l] Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster' tbl2 on node2 OK SHA512 Checking JSONExtractFloat('foo') Executing query select JSONExtractFloat('foo') on backward Stdout:0 Executing query CREATE DATABASE ordinary ENGINE=Ordinary on node Executing query select JSONExtractFloat('foo') on upstream Executing query CREATE DATABASE other ENGINE=Ordinary on node Executing query SELECT * FROM tbl2 on node2 OK JSONExtractFloat Checking JSONKey('foo') Executing query select JSONKey('foo') on backward Executing query CREATE DATABASE `.o r d i n a r y.` ENGINE=Ordinary on node Executing query DROP TABLE tbl2 ON CLUSTER 'cluster' SYNC on node1 Executing query select JSONKey('foo') on upstream Executing query CREATE DATABASE atomic ENGINE=Atomic on node OK JSONKey Checking JSONLength('foo') Executing query CREATE DATABASE mem ENGINE=Memory on node Executing query select JSONLength('foo') on backward Executing query CREATE DATABASE lazy ENGINE=Lazy(1) on node Executing query REVOKE ALL FROM u1 on node1 Executing query select JSONLength('foo') on upstream Executing query CREATE TABLE ordinary.t1 (n int) ENGINE=Memory on node Executing query RESTORE ALL ON CLUSTER 'cluster' FROM Disk('backups', '25') on node1 OK JSONLength Checking isValidJSON('foo') Executing query select isValidJSON('foo') on backward Executing query GRANT INSERT, CREATE TABLE ON tbl TO u1 on node1 Executing query CREATE TABLE ordinary.mt1 (n int) ENGINE=MergeTree order by n on node Executing query select isValidJSON('foo') on upstream Executing query RESTORE ALL ON CLUSTER 'cluster' FROM Disk('backups', '25') on node1 Executing query CREATE TABLE ordinary.mt2 (n int) ENGINE=MergeTree order by n on node OK isValidJSON Checking windowID('foo') Executing query select windowID('foo') on backward Executing query CREATE TABLE ordinary.rmt1 (n int, m int) ENGINE=ReplicatedMergeTree('/test/rmt1/ordinary', '1') order by n on node Skipping windowID Checking hopStart('foo') Executing query select hopStart('foo') on backward Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster' tbl on node2 Skipping hopStart Checking hop('foo') Executing query select hop('foo') on backward Executing query CREATE TABLE ordinary.rmt2 (n int, m int) ENGINE=ReplicatedMergeTree('/test/ordinary/rmt2', '1') order by n on node Skipping hop Checking detectTonality('foo') Executing query select detectTonality('foo') on backward Executing query SELECT * FROM tbl on node2 run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', "sed --follow-symlinks -i 's|/test/ordinary/rmt2|/test/{database}/{table}|' /var/lib/clickhouse/metadata/ordinary/rmt2.sql"] Command:[docker exec roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c sed --follow-symlinks -i 's|/test/ordinary/rmt2|/test/{database}/{table}|' /var/lib/clickhouse/metadata/ordinary/rmt2.sql] Skipping detectTonality Checking toRelativeWeekNum('foo') Executing query select toRelativeWeekNum('foo') on backward run container_id:roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query CREATE MATERIALIZED VIEW ordinary.mv1 (n int) ENGINE=ReplicatedMergeTree('/test/ordinary/mv1/', '1') order by n AS SELECT n FROM ordinary.rmt1 on node Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw0] PASSED test_backup_restore_on_cluster/test.py::test_required_privileges Skipping toRelativeWeekNum Checking makeDateTime64('foo') Executing query select makeDateTime64('foo') on backward Stdout:10 Skipping makeDateTime64 Checking makeDateTime('foo') Executing query select makeDateTime('foo') on backward Executing query CREATE MATERIALIZED VIEW ordinary.mv2 (n int) ENGINE=MergeTree order by n AS SELECT n FROM ordinary.rmt2 on node Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Skipping makeDateTime Checking makeDate('foo') Executing query select makeDate('foo') on backward Executing query CREATE DICTIONARY ordinary.d1 (n int DEFAULT 0, m int DEFAULT 1) PRIMARY KEY n SOURCE(CLICKHOUSE(HOST 'localhost' PORT 9000 USER 'default' TABLE 'rmt1' PASSWORD '' DB 'ordinary')) LIFETIME(MIN 1 MAX 10) LAYOUT(FLAT()) on node Skipping makeDate Checking transactionLatestSnapshot('foo') Executing query select transactionLatestSnapshot('foo') on backward Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Executing query CREATE DICTIONARY ordinary.d2 (n int DEFAULT 0, m int DEFAULT 1) PRIMARY KEY n SOURCE(CLICKHOUSE(HOST 'localhost' PORT 9000 USER 'default' TABLE 'rmt2' PASSWORD '' DB 'ordinary')) LIFETIME(MIN 1 MAX 10) LAYOUT(FLAT()) on node Stderr: node1 Skipped - Image is already being pulled by zoo3 Stderr: zoo1 Skipped - Image is already being pulled by zoo3 Stderr: zoo2 Skipped - Image is already being pulled by zoo3 Stderr: zoo3 Pulling Stderr: zoo3 Pulled Setup ZooKeeper Creating internal ZooKeeper dirs: ['/ClickHouse/tests/integration/test_broken_part_during_merge/_instances-0-gw8/keeper1/log', '/ClickHouse/tests/integration/test_broken_part_during_merge/_instances-0-gw8/keeper1/config', '/ClickHouse/tests/integration/test_broken_part_during_merge/_instances-0-gw8/keeper1/coordination', '/ClickHouse/tests/integration/test_broken_part_during_merge/_instances-0-gw8/keeper2/log', '/ClickHouse/tests/integration/test_broken_part_during_merge/_instances-0-gw8/keeper2/config', '/ClickHouse/tests/integration/test_broken_part_during_merge/_instances-0-gw8/keeper2/coordination', '/ClickHouse/tests/integration/test_broken_part_during_merge/_instances-0-gw8/keeper3/log', '/ClickHouse/tests/integration/test_broken_part_during_merge/_instances-0-gw8/keeper3/config', '/ClickHouse/tests/integration/test_broken_part_during_merge/_instances-0-gw8/keeper3/coordination'] Command:[docker compose --project-name roottestbrokenpartduringmerge-gw8 --env-file /ClickHouse/tests/integration/test_broken_part_during_merge/_instances-0-gw8/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --verbose up -d] Skipping transactionLatestSnapshot Checking transactionID('foo') Executing query select transactionID('foo') on backward Executing query CREATE TABLE ordinary.merge (n int) ENGINE=Merge('ordinary', '(mt)|(mv)') on node Skipping transactionID Checking extractGroups('foo') Executing query select extractGroups('foo') on backward Executing query CREATE TABLE ordinary.detached (n int) ENGINE=Log on node Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Stderr: zoo1 Skipped - Image is already being pulled by node1 Stderr: zoo2 Skipped - Image is already being pulled by node1 Stderr: zoo3 Skipped - Image is already being pulled by node1 Stderr: node1 Pulling Stderr: node1 Pulled Setup ZooKeeper Creating internal ZooKeeper dirs: ['/ClickHouse/tests/integration/test_cleanup_after_start/_instances-0-gw6/keeper1/log', '/ClickHouse/tests/integration/test_cleanup_after_start/_instances-0-gw6/keeper1/config', '/ClickHouse/tests/integration/test_cleanup_after_start/_instances-0-gw6/keeper1/coordination', '/ClickHouse/tests/integration/test_cleanup_after_start/_instances-0-gw6/keeper2/log', '/ClickHouse/tests/integration/test_cleanup_after_start/_instances-0-gw6/keeper2/config', '/ClickHouse/tests/integration/test_cleanup_after_start/_instances-0-gw6/keeper2/coordination', '/ClickHouse/tests/integration/test_cleanup_after_start/_instances-0-gw6/keeper3/log', '/ClickHouse/tests/integration/test_cleanup_after_start/_instances-0-gw6/keeper3/config', '/ClickHouse/tests/integration/test_cleanup_after_start/_instances-0-gw6/keeper3/coordination'] Command:[docker compose --project-name roottestcleanupafterstart-gw6 --env-file /ClickHouse/tests/integration/test_cleanup_after_start/_instances-0-gw6/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --verbose up -d] Skipping extractGroups Checking __bitWrapperFunc('foo') Executing query select __bitWrapperFunc('foo') on backward Executing query INSERT INTO ordinary.mt1 (n) VALUES (8) on node Skipping __bitWrapperFunc Checking showCertificate('foo') Executing query select showCertificate('foo') on backward Executing query INSERT INTO ordinary.mt2 (n) VALUES (8) on node Skipping showCertificate Checking nested('foo') Executing query select nested('foo') on backward Stderr: node Pulling Stderr: node Pulled ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_build_sets_from_multiple_threads/_instances-0-gw3/.env --project-name roottestbuildsetsfrommultiplethreads-gw3 --file /ClickHouse/tests/integration/test_build_sets_from_multiple_threads/_instances-0-gw3/node/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_build_sets_from_multiple_threads/_instances-0-gw3/.env --project-name roottestbuildsetsfrommultiplethreads-gw3 --file /ClickHouse/tests/integration/test_build_sets_from_multiple_threads/_instances-0-gw3/node/docker-compose.yml up -d --no-recreate] Executing query INSERT INTO ordinary.rmt1 (n) VALUES (8) on node Skipping nested Checking hashid('foo') Executing query select hashid('foo') on backward Skipping hashid Checking h3EdgeLengthM('foo') Executing query select h3EdgeLengthM('foo') on backward Executing query INSERT INTO ordinary.rmt2 (n) VALUES (8) on node Skipping h3EdgeLengthM Checking UUIDNumToString('foo') Executing query select UUIDNumToString('foo') on backward Executing query INSERT INTO ordinary.mv1 (n) VALUES (8) on node Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Skipping UUIDNumToString Checking IPv4StringToNumOrDefault('foo') Executing query select IPv4StringToNumOrDefault('foo') on backward run container_id:roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stderr:time="2025-04-02T03:34:57Z" level=trace msg="Docker Desktop integration not enabled" Stderr: Network roottestbrokenpartduringmerge-gw8_default Creating Stderr: Network roottestbrokenpartduringmerge-gw8_default Created Stderr: Container roottestbrokenpartduringmerge-gw8-zoo3-1 Creating Stderr: Container roottestbrokenpartduringmerge-gw8-zoo1-1 Creating Stderr: Container roottestbrokenpartduringmerge-gw8-zoo2-1 Creating Stderr: Container roottestbrokenpartduringmerge-gw8-zoo2-1 Created Stderr: Container roottestbrokenpartduringmerge-gw8-zoo1-1 Created Stderr: Container roottestbrokenpartduringmerge-gw8-zoo3-1 Created Stderr: Container roottestbrokenpartduringmerge-gw8-zoo3-1 Starting Stderr: Container roottestbrokenpartduringmerge-gw8-zoo1-1 Starting Stderr: Container roottestbrokenpartduringmerge-gw8-zoo2-1 Starting Stderr: Container roottestbrokenpartduringmerge-gw8-zoo1-1 Started Stderr: Container roottestbrokenpartduringmerge-gw8-zoo2-1 Started Stderr: Container roottestbrokenpartduringmerge-gw8-zoo3-1 Started Stderr:time="2025-04-02T03:34:58Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T03:34:58Z" level=debug msg="otel error" error="" Wait ZooKeeper to start get_instance_ip instance_name=zoo1 http://localhost:None "GET /v1.46/containers/roottestbrokenpartduringmerge-gw8-zoo1-1/json HTTP/1.1" 200 None get_kazoo_client: zoo1, ip:172.16.2.3, port:2181, use_ssl:False Connecting to 172.16.2.3(172.16.2.3):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Executing query INSERT INTO ordinary.mv2 (n) VALUES (8) on node Stdout:10 Executing query select IPv4StringToNumOrDefault('foo') on upstream test_backup_restore_on_cluster/test.py::test_shutdown_waits_for_backup Executing query CREATE TABLE tbl ON CLUSTER 'cluster' (x UInt8) ENGINE=ReplicatedMergeTree('/clickhouse/tables/tbl/', '{replica}')ORDER BY x on node1 Connecting to 172.16.2.3(172.16.2.3):2181, use_ssl: False Connection dropped: socket connection error: Connection refused OK IPv4StringToNumOrDefault Checking IPv4NumToStringClassC('foo') Executing query select IPv4NumToStringClassC('foo') on backward Executing query INSERT INTO ordinary.detached (n) VALUES (8) on node Executing query INSERT INTO tbl VALUES (3) on node1 Stderr:time="2025-04-02T03:34:57Z" level=trace msg="Docker Desktop integration not enabled" Stderr: Network roottestcleanupafterstart-gw6_default Creating Stderr: Network roottestcleanupafterstart-gw6_default Created Stderr: Container roottestcleanupafterstart-gw6-zoo1-1 Creating Stderr: Container roottestcleanupafterstart-gw6-zoo2-1 Creating Stderr: Container roottestcleanupafterstart-gw6-zoo3-1 Creating Stderr: Container roottestcleanupafterstart-gw6-zoo2-1 Created Stderr: Container roottestcleanupafterstart-gw6-zoo3-1 Created Stderr: Container roottestcleanupafterstart-gw6-zoo1-1 Created Stderr: Container roottestcleanupafterstart-gw6-zoo2-1 Starting Stderr: Container roottestcleanupafterstart-gw6-zoo3-1 Starting Stderr: Container roottestcleanupafterstart-gw6-zoo1-1 Starting Stderr: Container roottestcleanupafterstart-gw6-zoo2-1 Started Stderr: Container roottestcleanupafterstart-gw6-zoo3-1 Started Stderr: Container roottestcleanupafterstart-gw6-zoo1-1 Started Stderr:time="2025-04-02T03:34:58Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T03:34:58Z" level=debug msg="otel error" error="" Wait ZooKeeper to start get_instance_ip instance_name=zoo1 http://localhost:None "GET /v1.46/containers/roottestcleanupafterstart-gw6-zoo1-1/json HTTP/1.1" 200 None get_kazoo_client: zoo1, ip:172.16.6.4, port:2181, use_ssl:False Connecting to 172.16.6.4(172.16.6.4):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Executing query CREATE TABLE other.t1 (n int) ENGINE=Memory on node Connecting to 172.16.2.3(172.16.2.3):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Executing query INSERT INTO tbl VALUES (5) on node2 Skipping IPv4NumToStringClassC Checking IPv4NumToString('foo') Executing query select IPv4NumToString('foo') on backward Stderr: Network roottestbuildsetsfrommultiplethreads-gw3_default Creating Stderr: Network roottestbuildsetsfrommultiplethreads-gw3_default Created Stderr: Container roottestbuildsetsfrommultiplethreads-gw3-node-1 Creating Stderr: Container roottestbuildsetsfrommultiplethreads-gw3-node-1 Created Stderr: Container roottestbuildsetsfrommultiplethreads-gw3-node-1 Starting Stderr: Container roottestbuildsetsfrommultiplethreads-gw3-node-1 Started ClickHouse instance created get_instance_ip instance_name=node http://localhost:None "GET /v1.46/containers/roottestbuildsetsfrommultiplethreads-gw3-node-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node http://localhost:None "GET /v1.46/containers/roottestbuildsetsfrommultiplethreads-gw3-node-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node, ip: 172.16.7.2... http://localhost:None "GET /v1.46/containers/roottestbuildsetsfrommultiplethreads-gw3-node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/ddb363528299a161c4dea11a904622803edc67ee3e516047ea95c1679ad7d730/json HTTP/1.1" 200 None Executing query CREATE TABLE other.mt1 (n int) ENGINE=MergeTree order by n on node Executing query BACKUP TABLE tbl ON CLUSTER 'cluster' TO Disk('backups', '26') ASYNC on node1 Skipping IPv4NumToString Checking isIPv6String('foo') Executing query select isIPv6String('foo') on backward Connecting to 172.16.6.4(172.16.6.4):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Executing query CREATE TABLE other.mt2 (n int) ENGINE=MergeTree order by n on node run container_id:roottestbackuprestoreoncluster-gw0-node2-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] Command:[docker exec -u root roottestbackuprestoreoncluster-gw0-node2-1 bash -c ps -C clickhouse] Executing query select isIPv6String('foo') on upstream http://localhost:None "GET /v1.46/containers/ddb363528299a161c4dea11a904622803edc67ee3e516047ea95c1679ad7d730/json HTTP/1.1" 200 None Stdout: PID TTY TIME CMD Stdout: 10 ? 00:00:06 clickhouse run container_id:roottestbackuprestoreoncluster-gw0-node2-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] Command:[docker exec -u root roottestbackuprestoreoncluster-gw0-node2-1 bash -c pkill clickhouse] Executing query CREATE TABLE other.rmt1 (n int, m int) ENGINE=ReplicatedMergeTree('/test/rmt1/other', '1') order by n on node OK isIPv6String Checking JSON_QUERY('foo') Executing query select JSON_QUERY('foo') on backward run container_id:roottestbackuprestoreoncluster-gw0-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestoreoncluster-gw0-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query CREATE TABLE other.rmt2 (n int, m int) ENGINE=ReplicatedMergeTree('/test/other/rmt2', '1') order by n on node http://localhost:None "GET /v1.46/containers/ddb363528299a161c4dea11a904622803edc67ee3e516047ea95c1679ad7d730/json HTTP/1.1" 200 None Stdout:10 Skipping JSON_QUERY Checking isIPv4String('foo') Executing query select isIPv4String('foo') on backward run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', "sed --follow-symlinks -i 's|/test/other/rmt2|/test/{database}/{table}|' /var/lib/clickhouse/metadata/other/rmt2.sql"] Command:[docker exec roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c sed --follow-symlinks -i 's|/test/other/rmt2|/test/{database}/{table}|' /var/lib/clickhouse/metadata/other/rmt2.sql] Connecting to 172.16.6.4(172.16.6.4):2181, use_ssl: False Connection dropped: socket connection error: Connection refused http://localhost:None "GET /v1.46/containers/ddb363528299a161c4dea11a904622803edc67ee3e516047ea95c1679ad7d730/json HTTP/1.1" 200 None Executing query CREATE MATERIALIZED VIEW other.mv1 (n int) ENGINE=ReplicatedMergeTree('/test/other/mv1/', '1') order by n AS SELECT n FROM other.rmt1 on node Executing query select isIPv4String('foo') on upstream OK isIPv4String Checking IPv4CIDRToRange('foo') Executing query select IPv4CIDRToRange('foo') on backward http://localhost:None "GET /v1.46/containers/ddb363528299a161c4dea11a904622803edc67ee3e516047ea95c1679ad7d730/json HTTP/1.1" 200 None ClickHouse node started Executing query CREATE TABLE 02581_trips (id UInt32, description String, id2 UInt32) ENGINE = MergeTree PRIMARY KEY id ORDER BY id; INSERT INTO 02581_trips SELECT number, '', number FROM numbers(1); INSERT INTO 02581_trips SELECT number, '', number FROM numbers(1); INSERT INTO 02581_trips SELECT number, '', number FROM numbers(1); INSERT INTO 02581_trips SELECT number, '', number FROM numbers(1); INSERT INTO 02581_trips SELECT number, '', number FROM numbers(1); INSERT INTO 02581_trips SELECT number, '', number FROM numbers(1); INSERT INTO 02581_trips SELECT number, '', number FROM numbers(1); INSERT INTO 02581_trips SELECT number, '', number FROM numbers(1); INSERT INTO 02581_trips SELECT number, '', number FROM numbers(1); INSERT INTO 02581_trips SELECT number, '', number FROM numbers(1); INSERT INTO 02581_trips SELECT number, '', number FROM numbers(1); INSERT INTO 02581_trips SELECT number, '', number FROM numbers(1); INSERT INTO 02581_trips on node Executing query CREATE MATERIALIZED VIEW other.mv2 (n int) ENGINE=MergeTree order by n AS SELECT n FROM other.rmt2 on node Skipping IPv4CIDRToRange Checking IPv6CIDRToRange('foo') Executing query select IPv6CIDRToRange('foo') on backward Executing query CREATE DICTIONARY other.d1 (n int DEFAULT 0, m int DEFAULT 1) PRIMARY KEY n SOURCE(CLICKHOUSE(HOST 'localhost' PORT 9000 USER 'default' TABLE 'rmt1' PASSWORD '' DB 'other')) LIFETIME(MIN 1 MAX 10) LAYOUT(FLAT()) on node Skipping IPv6CIDRToRange Checking MACStringToNum('foo') Executing query select MACStringToNum('foo') on backward Connecting to 172.16.2.3(172.16.2.3):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Executing query CREATE DICTIONARY other.d2 (n int DEFAULT 0, m int DEFAULT 1) PRIMARY KEY n SOURCE(CLICKHOUSE(HOST 'localhost' PORT 9000 USER 'default' TABLE 'rmt2' PASSWORD '' DB 'other')) LIFETIME(MIN 1 MAX 10) LAYOUT(FLAT()) on node Executing query select MACStringToNum('foo') on upstream Executing query CREATE TABLE other.merge (n int) ENGINE=Merge('other', '(mt)|(mv)') on node run container_id:roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query ALTER TABLE `02581_trips` UPDATE description = 'a' WHERE id IN (SELECT CAST(number * 10, 'UInt32') FROM numbers(10e9)) SETTINGS mutations_sync = 2 on node OK MACStringToNum Checking cutIPv6('foo') Executing query select cutIPv6('foo') on backward Stdout:10 Executing query CREATE TABLE other.detached (n int) ENGINE=Log on node Skipping cutIPv6 Checking snowflakeToDateTime64('foo') Executing query select snowflakeToDateTime64('foo') on backward Connecting to 172.16.6.4(172.16.6.4):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Executing query INSERT INTO other.mt1 (n) VALUES (5) on node Skipping snowflakeToDateTime64 Checking radians('foo') Executing query select radians('foo') on backward Executing query INSERT INTO other.mt2 (n) VALUES (5) on node Skipping radians Checking ifNotFinite('foo') Executing query select ifNotFinite('foo') on backward Executing query INSERT INTO other.rmt1 (n) VALUES (5) on node Skipping ifNotFinite Checking polygonAreaSpherical('foo') Executing query select polygonAreaSpherical('foo') on backward Executing query INSERT INTO other.rmt2 (n) VALUES (5) on node Skipping polygonAreaSpherical Checking base58Encode('foo') Executing query select base58Encode('foo') on backward Executing query INSERT INTO other.mv1 (n) VALUES (5) on node Executing query select base58Encode('foo') on upstream Executing query INSERT INTO other.mv2 (n) VALUES (5) on node OK base58Encode Checking currentUser('foo') Executing query select currentUser('foo') on backward Skipping currentUser Checking addSeconds('foo') Executing query select addSeconds('foo') on backward Executing query INSERT INTO other.detached (n) VALUES (5) on node Skipping addSeconds Checking convertCharset('foo') Executing query select convertCharset('foo') on backward Skipping convertCharset Checking extractTextFromHTML('foo') Executing query select extractTextFromHTML('foo') on backward Executing query CREATE TABLE atomic.t1 (n int) ENGINE=Memory on node run container_id:roottestbackuprestoreoncluster-gw0-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestoreoncluster-gw0-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select extractTextFromHTML('foo') on upstream Stdout:10 Executing query CREATE TABLE atomic.mt1 (n int) ENGINE=MergeTree order by n on node OK extractTextFromHTML Checking toIntervalQuarter('foo') Executing query select toIntervalQuarter('foo') on backward Executing query CREATE TABLE atomic.mt2 (n int) ENGINE=MergeTree order by n on node Skipping toIntervalQuarter Checking multiSearchAny('foo') Executing query select multiSearchAny('foo') on backward Executing query CREATE TABLE atomic.rmt1 (n int, m int) ENGINE=ReplicatedMergeTree('/test/rmt1/atomic', '1') order by n on node Skipping multiSearchAny Checking bitmapContains('foo') Executing query select bitmapContains('foo') on backward Executing query CREATE TABLE atomic.rmt2 (n int, m int) ENGINE=ReplicatedMergeTree('/test/atomic/rmt2', '1') order by n on node Skipping bitmapContains Checking bitmapHasAny('foo') Executing query select bitmapHasAny('foo') on backward run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', "sed --follow-symlinks -i 's|/test/atomic/rmt2|/test/{database}/{table}|' /var/lib/clickhouse/metadata/atomic/rmt2.sql"] Command:[docker exec roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c sed --follow-symlinks -i 's|/test/atomic/rmt2|/test/{database}/{table}|' /var/lib/clickhouse/metadata/atomic/rmt2.sql] Executing query CREATE MATERIALIZED VIEW atomic.mv1 (n int) ENGINE=ReplicatedMergeTree('/test/atomic/mv1/', '1') order by n AS SELECT n FROM atomic.rmt1 on node Skipping bitmapHasAny Checking bitmapXor('foo') Executing query select bitmapXor('foo') on backward Executing query CREATE MATERIALIZED VIEW atomic.mv2 (n int) ENGINE=MergeTree order by n AS SELECT n FROM atomic.rmt2 on node Skipping bitmapXor Checking bitmapOr('foo') Executing query select bitmapOr('foo') on backward run container_id:roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query CREATE DICTIONARY atomic.d1 (n int DEFAULT 0, m int DEFAULT 1) PRIMARY KEY n SOURCE(CLICKHOUSE(HOST 'localhost' PORT 9000 USER 'default' TABLE 'rmt1' PASSWORD '' DB 'atomic')) LIFETIME(MIN 1 MAX 10) LAYOUT(FLAT()) on node Stderr: Container roottestclusterallreplicas-gw1-node2-1 Stopping Stderr: Container roottestclusterallreplicas-gw1-node1-1 Stopping Stderr: Container roottestclusterallreplicas-gw1-node1-1 Stopped Stderr: Container roottestclusterallreplicas-gw1-node2-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw1/node1/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw1/node1/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Skipping bitmapOr Checking xxh3('foo') Executing query select xxh3('foo') on backward Command:[bash -c [ -f /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw1/node2/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw1/node2/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Stdout:10 Command:[docker compose --env-file /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw1/.env --project-name roottestclusterallreplicas-gw1 --file /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw1/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw1/node2/docker-compose.yml down --volumes] Connecting to 172.16.2.3(172.16.2.3):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Executing query CREATE DICTIONARY atomic.d2 (n int DEFAULT 0, m int DEFAULT 1) PRIMARY KEY n SOURCE(CLICKHOUSE(HOST 'localhost' PORT 9000 USER 'default' TABLE 'rmt2' PASSWORD '' DB 'atomic')) LIFETIME(MIN 1 MAX 10) LAYOUT(FLAT()) on node Executing query select xxh3('foo') on upstream Executing query CREATE TABLE atomic.merge (n int) ENGINE=Merge('atomic', '(mt)|(mv)') on node Connecting to 172.16.6.4(172.16.6.4):2181, use_ssl: False Connection dropped: socket connection error: Connection refused OK xxh3 Checking bitmapAnd('foo') Executing query select bitmapAnd('foo') on backward Executing query CREATE TABLE atomic.detached (n int) ENGINE=Log on node Skipping bitmapAnd Checking bitmapAndnotCardinality('foo') Executing query select bitmapAndnotCardinality('foo') on backward Executing query INSERT INTO atomic.mt1 (n) VALUES (6) on node Skipping bitmapAndnotCardinality Checking IPv6StringToNumOrDefault('foo') Executing query select IPv6StringToNumOrDefault('foo') on backward Executing query INSERT INTO atomic.mt2 (n) VALUES (6) on node Executing query select IPv6StringToNumOrDefault('foo') on upstream OK IPv6StringToNumOrDefault Checking bitmapAndnot('foo') Executing query select bitmapAndnot('foo') on backward Executing query INSERT INTO atomic.rmt1 (n) VALUES (6) on node Skipping bitmapAndnot Checking bitmapXorCardinality('foo') Executing query select bitmapXorCardinality('foo') on backward Executing query INSERT INTO atomic.rmt2 (n) VALUES (6) on node run container_id:roottestbackuprestoreoncluster-gw0-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestoreoncluster-gw0-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Skipping bitmapXorCardinality Checking bitmapOrCardinality('foo') Executing query select bitmapOrCardinality('foo') on backward Stdout:10 Executing query INSERT INTO atomic.mv1 (n) VALUES (6) on node Skipping bitmapOrCardinality Checking bitmapAndCardinality('foo') Executing query select bitmapAndCardinality('foo') on backward Stderr: Container roottestclusterallreplicas-gw1-node1-1 Stopping Stderr: Container roottestclusterallreplicas-gw1-node2-1 Stopping Stderr: Container roottestclusterallreplicas-gw1-node1-1 Stopped Stderr: Container roottestclusterallreplicas-gw1-node1-1 Removing Stderr: Container roottestclusterallreplicas-gw1-node2-1 Stopped Stderr: Container roottestclusterallreplicas-gw1-node2-1 Removing Stderr: Container roottestclusterallreplicas-gw1-node2-1 Removed Stderr: Container roottestclusterallreplicas-gw1-node1-1 Removed Stderr: Network roottestclusterallreplicas-gw1_default Removing Stderr: Network roottestclusterallreplicas-gw1_default Removed Cleanup called Docker networks for project roottestclusterallreplicas-gw1 are NETWORK ID NAME DRIVER SCOPE Executing query INSERT INTO atomic.mv2 (n) VALUES (6) on node Docker containers for project roottestclusterallreplicas-gw1 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Skipping bitmapAndCardinality Checking bitmapMin('foo') Executing query select bitmapMin('foo') on backward Docker volumes for project roottestclusterallreplicas-gw1 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestclusterallreplicas-gw1-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestclusterallreplicas-gw1 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Executing query INSERT INTO atomic.detached (n) VALUES (6) on node Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Skipping bitmapMin Checking bitmapCardinality('foo') Executing query select bitmapCardinality('foo') on backward Stdout:1 Volumes pruned: 1 test_backward_compatibility/test_vertical_merges_from_compact_parts.py::test_vertical_merges_from_compact_parts Running tests in /ClickHouse/tests/integration/test_backward_compatibility/test_vertical_merges_from_compact_parts.py Cluster start called. is_up=False Docker networks for project roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Executing query CREATE TABLE `.o r d i n a r y.`.`t. a. b. l. e.` (n int) ENGINE=MergeTree ORDER BY n on node Docker volumes for project roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1 are DRIVER VOLUME NAME Cleanup called Skipping bitmapCardinality Docker networks for project roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1 are NETWORK ID NAME DRIVER SCOPE Checking bitmapSubsetLimit('foo') Executing query select bitmapSubsetLimit('foo') on backward Docker containers for project roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Executing query CREATE TABLE lazy.table (n int) ENGINE=Log on node Docker volumes for project roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1 Trying to prune unused networks... Skipping bitmapSubsetLimit Checking bitmapSubsetInRange('foo') Executing query select bitmapSubsetInRange('foo') on backward Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Executing query CREATE TABLE ordinary.l (n DEFAULT dictGet('other.d1', 'm', toUInt64(3))) ENGINE=Log on node Stdout:1 Volumes pruned: 1 Setup directory for instance: node1 Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files [] to /ClickHouse/tests/integration/test_backward_compatibility/_instances-vertical_merges_from_compact_parts-0-gw1/node1/configs/config.d Setup database dir /ClickHouse/tests/integration/test_backward_compatibility/_instances-vertical_merges_from_compact_parts-0-gw1/node1/database Setup logs dir /ClickHouse/tests/integration/test_backward_compatibility/_instances-vertical_merges_from_compact_parts-0-gw1/node1/logs Entrypoint cmd: bash -c "trap 'pkill tail' INT TERM; clickhouse server --config-file=/etc/clickhouse-server/config.xml --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log --daemon -- ; coproc tail -f /dev/null; wait $$!" Setup directory for instance: node2 Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_backward_compatibility/configs/no_compress_marks.xml', '/ClickHouse/tests/integration/test_backward_compatibility/configs/no_allow_vertical_merges_from_compact_to_wide_parts.xml'] to /ClickHouse/tests/integration/test_backward_compatibility/_instances-vertical_merges_from_compact_parts-0-gw1/node2/configs/config.d Setup database dir /ClickHouse/tests/integration/test_backward_compatibility/_instances-vertical_merges_from_compact_parts-0-gw1/node2/database Setup logs dir /ClickHouse/tests/integration/test_backward_compatibility/_instances-vertical_merges_from_compact_parts-0-gw1/node2/logs Entrypoint cmd: bash -c "trap 'pkill tail' INT TERM; clickhouse server --config-file=/etc/clickhouse-server/config.xml --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log --daemon -- ; coproc tail -f /dev/null; wait $$!" Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'keeper_binary': '/clickhouse', 'keeper_cmd_prefix': 'clickhouse keeper', 'image': 'altinityinfra/integration-test:8b2301119731', 'user': '0', 'keeper_fs': 'bind', 'keeper_logs_dir1': '/ClickHouse/tests/integration/test_backward_compatibility/_instances-vertical_merges_from_compact_parts-0-gw1/keeper1/log', 'keeper_config_dir1': '/ClickHouse/tests/integration/test_backward_compatibility/_instances-vertical_merges_from_compact_parts-0-gw1/keeper1/config', 'keeper_db_dir1': '/ClickHouse/tests/integration/test_backward_compatibility/_instances-vertical_merges_from_compact_parts-0-gw1/keeper1/coordination', 'keeper_logs_dir2': '/ClickHouse/tests/integration/test_backward_compatibility/_instances-vertical_merges_from_compact_parts-0-gw1/keeper2/log', 'keeper_config_dir2': '/ClickHouse/tests/integration/test_backward_compatibility/_instances-vertical_merges_from_compact_parts-0-gw1/keeper2/config', 'keeper_db_dir2': '/ClickHouse/tests/integration/test_backward_compatibility/_instances-vertical_merges_from_compact_parts-0-gw1/keeper2/coordination', 'keeper_logs_dir3': '/ClickHouse/tests/integration/test_backward_compatibility/_instances-vertical_merges_from_compact_parts-0-gw1/keeper3/log', 'keeper_config_dir3': '/ClickHouse/tests/integration/test_backward_compatibility/_instances-vertical_merges_from_compact_parts-0-gw1/keeper3/config', 'keeper_db_dir3': '/ClickHouse/tests/integration/test_backward_compatibility/_instances-vertical_merges_from_compact_parts-0-gw1/keeper3/coordination'} stored in /ClickHouse/tests/integration/test_backward_compatibility/_instances-vertical_merges_from_compact_parts-0-gw1/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_backward_compatibility/_instances-vertical_merges_from_compact_parts-0-gw1/.env --project-name roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1 --file /ClickHouse/tests/integration/test_backward_compatibility/_instances-vertical_merges_from_compact_parts-0-gw1/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_backward_compatibility/_instances-vertical_merges_from_compact_parts-0-gw1/node2/docker-compose.yml pull] Skipping bitmapSubsetInRange Checking bitmapToArray('foo') Executing query select bitmapToArray('foo') on backward Executing query CREATE TABLE other.l (n DEFAULT dictGet('ordinary.d1', 'm', toUInt64(3))) ENGINE=StripeLog on node run container_id:roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Skipping bitmapToArray Checking bitmapBuild('foo') Executing query select bitmapBuild('foo') on backward Executing query CREATE TABLE atomic.l (n DEFAULT dictGet('ordinary.d1', 'm', toUInt64(3))) ENGINE=TinyLog on node Stdout:10 Skipping bitmapBuild Checking bitHammingDistance('foo') Executing query select bitHammingDistance('foo') on backward Skipping bitHammingDistance Checking polygonsDistanceSpherical('foo') Executing query select polygonsDistanceSpherical('foo') on backward Executing query SELECT count(), sum(n) FROM ordinary.merge on node Skipping polygonsDistanceSpherical Checking polygonsDistanceCartesian('foo') Executing query select polygonsDistanceCartesian('foo') on backward Executing query SELECT count(), sum(n) FROM other.merge on node Skipping polygonsDistanceCartesian Checking h3EdgeLengthKm('foo') Executing query select h3EdgeLengthKm('foo') on backward Skipping h3EdgeLengthKm Checking fuzzBits('foo') Executing query select fuzzBits('foo') on backward Executing query SELECT count(), sum(n) FROM atomic.merge on node Skipping fuzzBits Checking addressToLineWithInlines('foo') Executing query select addressToLineWithInlines('foo') on backward Skipping addressToLineWithInlines Checking monthName('foo') Executing query select monthName('foo') on backward Executing query DETACH TABLE ordinary.detached PERMANENTLY on node Skipping monthName Checking arrayReverseFill('foo') Executing query select arrayReverseFill('foo') on backward run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', 'touch /var/lib/clickhouse/flags/convert_ordinary_to_atomic'] Command:[docker exec roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c touch /var/lib/clickhouse/flags/convert_ordinary_to_atomic] run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] Command:[docker exec -u root roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c ps -C clickhouse] Skipping arrayReverseFill Checking arrayFill('foo') Executing query select arrayFill('foo') on backward Stdout: PID TTY TIME CMD Stdout: 401 ? 00:00:01 clickhouse run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] Command:[docker exec -u root roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c pkill clickhouse] run container_id:roottestbackuprestoreoncluster-gw0-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestoreoncluster-gw0-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Skipping arrayFill Checking addressToSymbol('foo') Executing query select addressToSymbol('foo') on backward run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:10 Stdout:401 Skipping addressToSymbol Checking aes_decrypt_mysql('foo') Executing query select aes_decrypt_mysql('foo') on backward Skipping aes_decrypt_mysql Checking pathFull('foo') Executing query select pathFull('foo') on backward Executing query select pathFull('foo') on upstream Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=38, time_out=30000, session_id=1, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=10, time_out=30000, session_id=2, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Zookeeper connection established, state: CONNECTED Connecting to 172.16.2.3(172.16.2.3):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost OK pathFull Checking dictGetDescendants('foo') Executing query select dictGetDescendants('foo') on backward Skipping dictGetDescendants Checking dictGetChildren('foo') Executing query select dictGetChildren('foo') on backward Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED get_instance_ip instance_name=zoo2 http://localhost:None "GET /v1.46/containers/roottestbrokenpartduringmerge-gw8-zoo2-1/json HTTP/1.1" 200 None get_kazoo_client: zoo2, ip:172.16.2.2, port:2181, use_ssl:False Connecting to 172.16.2.2(172.16.2.2):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost Command:[docker compose --env-file /ClickHouse/tests/integration/test_build_sets_from_multiple_threads/_instances-0-gw3/.env --project-name roottestbuildsetsfrommultiplethreads-gw3 --file /ClickHouse/tests/integration/test_build_sets_from_multiple_threads/_instances-0-gw3/node/docker-compose.yml stop --timeout 20] [gw3] PASSED test_build_sets_from_multiple_threads/test.py::test_set Skipping dictGetChildren Checking notEmpty('foo') Executing query select notEmpty('foo') on backward Executing query select notEmpty('foo') on upstream run container_id:roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED get_instance_ip instance_name=zoo3 http://localhost:None "GET /v1.46/containers/roottestbrokenpartduringmerge-gw8-zoo3-1/json HTTP/1.1" 200 None get_kazoo_client: zoo3, ip:172.16.2.4, port:2181, use_ssl:False Connecting to 172.16.2.4(172.16.2.4):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost Stdout:10 OK notEmpty Checking format('foo') Executing query select format('foo') on backward Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED All instances of ZooKeeper started: ('zoo1', 'zoo2', 'zoo3') ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_broken_part_during_merge/_instances-0-gw8/.env --project-name roottestbrokenpartduringmerge-gw8 --file /ClickHouse/tests/integration/test_broken_part_during_merge/_instances-0-gw8/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_broken_part_during_merge/_instances-0-gw8/.env --project-name roottestbrokenpartduringmerge-gw8 --file /ClickHouse/tests/integration/test_broken_part_during_merge/_instances-0-gw8/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml up -d --no-recreate] Skipping format Checking dictGetHierarchy('foo') Executing query select dictGetHierarchy('foo') on backward Skipping dictGetHierarchy Checking dictGetStringOrDefault('foo') Executing query select dictGetStringOrDefault('foo') on backward Skipping dictGetStringOrDefault Checking dictGetUUIDOrDefault('foo') Executing query select dictGetUUIDOrDefault('foo') on backward Skipping dictGetUUIDOrDefault Checking dictGetDateTimeOrDefault('foo') Executing query select dictGetDateTimeOrDefault('foo') on backward Skipping dictGetDateTimeOrDefault Checking UTCTimestamp('foo') Executing query select UTCTimestamp('foo') on backward Connecting to 172.16.6.4(172.16.6.4):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Stderr: Container roottestbrokenpartduringmerge-gw8-zoo1-1 Running Stderr: Container roottestbrokenpartduringmerge-gw8-zoo2-1 Running Stderr: Container roottestbrokenpartduringmerge-gw8-zoo3-1 Running Stderr: Container roottestbrokenpartduringmerge-gw8-node1-1 Creating Stderr: Container roottestbrokenpartduringmerge-gw8-node1-1 Created Stderr: Container roottestbrokenpartduringmerge-gw8-node1-1 Starting Stderr: Container roottestbrokenpartduringmerge-gw8-node1-1 Started ClickHouse instance created get_instance_ip instance_name=node1 Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Sending request(xid=2): Close() http://localhost:None "GET /v1.46/containers/roottestbrokenpartduringmerge-gw8-node1-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node1 Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost http://localhost:None "GET /v1.46/containers/roottestbrokenpartduringmerge-gw8-node1-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node1, ip: 172.16.2.5... http://localhost:None "GET /v1.46/containers/roottestbrokenpartduringmerge-gw8-node1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/60459101f2594f1df616748aa91373f01132d541daca91284beda9e2f00947ec/json HTTP/1.1" 200 None Skipping UTCTimestamp Checking dictGetFloat64OrDefault('foo') Executing query select dictGetFloat64OrDefault('foo') on backward Skipping dictGetFloat64OrDefault Checking dictGetInt64OrDefault('foo') Executing query select dictGetInt64OrDefault('foo') on backward run container_id:roottestbackuprestoreoncluster-gw0-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestoreoncluster-gw0-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED get_instance_ip instance_name=zoo2 http://localhost:None "GET /v1.46/containers/roottestcleanupafterstart-gw6-zoo2-1/json HTTP/1.1" 200 None get_kazoo_client: zoo2, ip:172.16.6.2, port:2181, use_ssl:False Connecting to 172.16.6.2(172.16.6.2):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED http://localhost:None "GET /v1.46/containers/60459101f2594f1df616748aa91373f01132d541daca91284beda9e2f00947ec/json HTTP/1.1" 200 None Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost Stdout:10 Skipping dictGetInt64OrDefault Checking today('foo') Executing query select today('foo') on backward Stdout:401 Skipping today Checking dictGetUInt32OrDefault('foo') Executing query select dictGetUInt32OrDefault('foo') on backward http://localhost:None "GET /v1.46/containers/60459101f2594f1df616748aa91373f01132d541daca91284beda9e2f00947ec/json HTTP/1.1" 200 None Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED get_instance_ip instance_name=zoo3 http://localhost:None "GET /v1.46/containers/roottestcleanupafterstart-gw6-zoo3-1/json HTTP/1.1" 200 None get_kazoo_client: zoo3, ip:172.16.6.3, port:2181, use_ssl:False Connecting to 172.16.6.3(172.16.6.3):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost Skipping dictGetUInt32OrDefault Checking h3IsPentagon('foo') Executing query select h3IsPentagon('foo') on backward http://localhost:None "GET /v1.46/containers/60459101f2594f1df616748aa91373f01132d541daca91284beda9e2f00947ec/json HTTP/1.1" 200 None Skipping h3IsPentagon Checking dictGetUInt16OrDefault('foo') Executing query select dictGetUInt16OrDefault('foo') on backward Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED All instances of ZooKeeper started: ('zoo1', 'zoo2', 'zoo3') ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_cleanup_after_start/_instances-0-gw6/.env --project-name roottestcleanupafterstart-gw6 --file /ClickHouse/tests/integration/test_cleanup_after_start/_instances-0-gw6/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_cleanup_after_start/_instances-0-gw6/.env --project-name roottestcleanupafterstart-gw6 --file /ClickHouse/tests/integration/test_cleanup_after_start/_instances-0-gw6/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml up -d --no-recreate] Skipping dictGetUInt16OrDefault Checking arrayIntersect('foo') Executing query select arrayIntersect('foo') on backward http://localhost:None "GET /v1.46/containers/60459101f2594f1df616748aa91373f01132d541daca91284beda9e2f00947ec/json HTTP/1.1" 200 None ClickHouse node1 started Executing query CREATE TABLE replicated_mt(date Date, id UInt32, value Int32) ENGINE = ReplicatedMergeTree('/clickhouse/tables/replicated_mt', 'node1') ORDER BY id SETTINGS cleanup_delay_period=1, cleanup_delay_period_random_add=1, cleanup_thread_preferred_points_per_iteration=0; on node1 Skipping arrayIntersect Checking dictGetUInt8OrDefault('foo') Executing query select dictGetUInt8OrDefault('foo') on backward Executing query SYSTEM STOP REPLICATION QUEUES replicated_mt on node1 Skipping dictGetUInt8OrDefault Checking dictGetString('foo') Executing query select dictGetString('foo') on backward Executing query INSERT INTO replicated_mt SELECT toDate('2019-10-01'), number, number * number FROM numbers (0, 100000) on node1 run container_id:roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Skipping dictGetString Checking dictGetIPv6('foo') Executing query select dictGetIPv6('foo') on backward Stdout:10 Skipping dictGetIPv6 Checking multiSearchFirstPosition('foo') Executing query select multiSearchFirstPosition('foo') on backward Stderr: Container roottestcleanupafterstart-gw6-zoo2-1 Running Stderr: Container roottestcleanupafterstart-gw6-zoo3-1 Running Stderr: Container roottestcleanupafterstart-gw6-zoo1-1 Running Stderr: Container roottestcleanupafterstart-gw6-node1-1 Creating Stderr: Container roottestcleanupafterstart-gw6-node1-1 Created Stderr: Container roottestcleanupafterstart-gw6-node1-1 Starting Stderr: Container roottestcleanupafterstart-gw6-node1-1 Started ClickHouse instance created get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottestcleanupafterstart-gw6-node1-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottestcleanupafterstart-gw6-node1-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node1, ip: 172.16.6.5... http://localhost:None "GET /v1.46/containers/roottestcleanupafterstart-gw6-node1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/89dbbfb44f143daf9deb159515e3317e18ef41ba1c0cd9cc927e5240bc540e29/json HTTP/1.1" 200 None Skipping multiSearchFirstPosition Checking dictGetDate('foo') Executing query select dictGetDate('foo') on backward Executing query INSERT INTO replicated_mt SELECT toDate('2019-10-01'), number, number * number FROM numbers (100000, 100000) on node1 http://localhost:None "GET /v1.46/containers/89dbbfb44f143daf9deb159515e3317e18ef41ba1c0cd9cc927e5240bc540e29/json HTTP/1.1" 200 None Skipping dictGetDate Checking dictGetInt16('foo') Executing query select dictGetInt16('foo') on backward Executing query INSERT INTO replicated_mt SELECT toDate('2019-10-01'), number, number * number FROM numbers (200000, 100000) on node1 Skipping dictGetInt16 Checking dictGetUInt64('foo') Executing query select dictGetUInt64('foo') on backward Executing query INSERT INTO replicated_mt SELECT toDate('2019-10-01'), number, number * number FROM numbers (300000, 100000) on node1 http://localhost:None "GET /v1.46/containers/89dbbfb44f143daf9deb159515e3317e18ef41ba1c0cd9cc927e5240bc540e29/json HTTP/1.1" 200 None Skipping dictGetUInt64 Checking dictGetUInt32('foo') Executing query select dictGetUInt32('foo') on backward Executing query SELECT COUNT() FROM system.parts WHERE table='replicated_mt' AND active=1 on node1 Skipping dictGetUInt32 Checking replaceOne('foo') Executing query select replaceOne('foo') on backward http://localhost:None "GET /v1.46/containers/89dbbfb44f143daf9deb159515e3317e18ef41ba1c0cd9cc927e5240bc540e29/json HTTP/1.1" 200 None Executing query SELECT path FROM system.parts WHERE table = 'replicated_mt' and name = 'all_3_3_0' on node1 Skipping replaceOne Checking emptyArrayInt16('foo') Executing query select emptyArrayInt16('foo') on backward http://localhost:None "GET /v1.46/containers/89dbbfb44f143daf9deb159515e3317e18ef41ba1c0cd9cc927e5240bc540e29/json HTTP/1.1" 200 None ClickHouse node1 started Executing query CREATE TABLE test_table(date Date, id UInt32, dummy UInt32) ENGINE = ReplicatedMergeTree('/clickhouse/tables/test_table', 'node1') PARTITION BY date ORDER BY id on node1 run container_id:roottestbrokenpartduringmerge-gw8-node1-1 detach:False nothrow:False cmd: ['bash', '-c', 'cd /var/lib/clickhouse/store/9c5/9c5d38e4-e01e-42b0-99db-48a7358d7d2d/all_3_3_0/ && ls *.bin | head -n 1'] Command:[docker exec roottestbrokenpartduringmerge-gw8-node1-1 bash -c cd /var/lib/clickhouse/store/9c5/9c5d38e4-e01e-42b0-99db-48a7358d7d2d/all_3_3_0/ && ls *.bin | head -n 1] Skipping emptyArrayInt16 Checking dictGetOrDefault('foo') Executing query select dictGetOrDefault('foo') on backward run container_id:roottestbackuprestoreoncluster-gw0-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestoreoncluster-gw0-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:data.bin run container_id:roottestbrokenpartduringmerge-gw8-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "cd /var/lib/clickhouse/store/9c5/9c5d38e4-e01e-42b0-99db-48a7358d7d2d/all_3_3_0/ && ls *.bin | head -n 1 | xargs -I{} sh -c 'truncate -s -1 $1' -- {}"] Command:[docker exec --privileged roottestbrokenpartduringmerge-gw8-node1-1 bash -c cd /var/lib/clickhouse/store/9c5/9c5d38e4-e01e-42b0-99db-48a7358d7d2d/all_3_3_0/ && ls *.bin | head -n 1 | xargs -I{} sh -c 'truncate -s -1 $1' -- {}] Skipping dictGetOrDefault Checking JSONExtractKeysAndValues('foo') Executing query select JSONExtractKeysAndValues('foo') on backward Stdout:10 Stdout:401 Executing query INSERT INTO test_table VALUES (toDate('2020-01-01'), 1, 10) on node1 Executing query CHECK TABLE replicated_mt on node1 Executing query OPTIMIZE TABLE replicated_mt FINAL on node1 Skipping JSONExtractKeysAndValues Checking replicate('foo') Executing query select replicate('foo') on backward Executing query SELECT count() FROM test_table on node1 Executing query SYSTEM START REPLICATION QUEUES replicated_mt on node1 Skipping replicate Checking translateUTF8('foo') Executing query select translateUTF8('foo') on backward Skipping translateUTF8 Checking arrayCumSumNonNegative('foo') Executing query select arrayCumSumNonNegative('foo') on backward run container_id:roottestcleanupafterstart-gw6-node1-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] Command:[docker exec -u root roottestcleanupafterstart-gw6-node1-1 bash -c ps -C clickhouse] Skipping arrayCumSumNonNegative Checking detectCharset('foo') Executing query select detectCharset('foo') on backward Stdout: PID TTY TIME CMD Stdout: 10 ? 00:00:00 clickhouse run container_id:roottestcleanupafterstart-gw6-node1-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] Command:[docker exec -u root roottestcleanupafterstart-gw6-node1-1 bash -c pkill clickhouse] Skipping detectCharset Checking reinterpretAsInt32('foo') Executing query select reinterpretAsInt32('foo') on backward run container_id:roottestcleanupafterstart-gw6-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestcleanupafterstart-gw6-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:10 Executing query select reinterpretAsInt32('foo') on upstream run container_id:roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] OK reinterpretAsInt32 Checking dictGetFloat64('foo') Executing query select dictGetFloat64('foo') on backward Stdout:10 Skipping dictGetFloat64 Checking arraySlice('foo') Executing query select arraySlice('foo') on backward Skipping arraySlice Checking toStartOfQuarter('foo') Executing query select toStartOfQuarter('foo') on backward Skipping toStartOfQuarter Checking timezoneOf('foo') Executing query select timezoneOf('foo') on backward Skipping timezoneOf Checking bitCount('foo') Executing query select bitCount('foo') on backward Skipping bitCount Checking atan2('foo') Executing query select atan2('foo') on backward Skipping atan2 Checking dictIsIn('foo') Executing query select dictIsIn('foo') on backward Skipping dictIsIn Checking base64Encode('foo') Executing query select base64Encode('foo') on backward Executing query select base64Encode('foo') on upstream run container_id:roottestbackuprestoreoncluster-gw0-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestoreoncluster-gw0-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:10 OK base64Encode Checking globalVariable('foo') Executing query select globalVariable('foo') on backward Stdout:401 Executing query select globalVariable('foo') on upstream OK globalVariable Checking divideDecimal('foo') Executing query select divideDecimal('foo') on backward Executing query ALTER TABLE replicated_mt UPDATE value = 7 WHERE 1 on node1 Skipping divideDecimal Checking fullHostName('foo') Executing query select fullHostName('foo') on backward Skipping fullHostName Checking arrayReduceInRanges('foo') Executing query select arrayReduceInRanges('foo') on backward Executing query SELECT sum(value) FROM replicated_mt on node1 Skipping arrayReduceInRanges Checking ignore('foo') Executing query select ignore('foo') on backward run container_id:roottestcleanupafterstart-gw6-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestcleanupafterstart-gw6-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query DROP TABLE replicated_mt SYNC on node1 Stdout:10 Executing query select ignore('foo') on upstream run container_id:roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker compose --env-file /ClickHouse/tests/integration/test_broken_part_during_merge/_instances-0-gw8/.env --project-name roottestbrokenpartduringmerge-gw8 --file /ClickHouse/tests/integration/test_broken_part_during_merge/_instances-0-gw8/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml stop --timeout 20] [gw8] PASSED test_broken_part_during_merge/test.py::test_merge_and_part_corruption OK ignore Checking normalizeQuery('foo') Executing query select normalizeQuery('foo') on backward Stdout:10 Executing query select normalizeQuery('foo') on upstream OK normalizeQuery Checking subtractMinutes('foo') Executing query select subtractMinutes('foo') on backward Skipping subtractMinutes Checking h3Line('foo') Executing query select h3Line('foo') on backward Skipping h3Line Checking bitTest('foo') Executing query select bitTest('foo') on backward Skipping bitTest Checking ngramSearchCaseInsensitiveUTF8('foo') Executing query select ngramSearchCaseInsensitiveUTF8('foo') on backward Skipping ngramSearchCaseInsensitiveUTF8 Checking ngramDistance('foo') Executing query select ngramDistance('foo') on backward Skipping ngramDistance Checking acos('foo') Executing query select acos('foo') on backward Skipping acos Checking accurateCastOrNull('foo') Executing query select accurateCastOrNull('foo') on backward run container_id:roottestbackuprestoreoncluster-gw0-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestoreoncluster-gw0-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:10 Skipping accurateCastOrNull Checking empty('foo') Executing query select empty('foo') on backward run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] No clickhouse process running. Start new one. http://localhost:None "POST /v1.46/containers/roottestbackwardcompatibilityconvertordinary-gw5-node-1/exec HTTP/1.1" 201 74 Executing query select empty('foo') on upstream http://localhost:None "POST /v1.46/exec/9e8d5dfb66f2715ef734dfbeee2301a5f376f58633dac235b0e89119038b8d75/start HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/exec/9e8d5dfb66f2715ef734dfbeee2301a5f376f58633dac235b0e89119038b8d75/json HTTP/1.1" 200 586 OK empty Checking ngramMinHashArg('foo') Executing query select ngramMinHashArg('foo') on backward Executing query select ngramMinHashArg('foo') on upstream OK ngramMinHashArg Checking _CAST('foo') Executing query select _CAST('foo') on backward Skipping _CAST Checking toDayOfYear('foo') Executing query select toDayOfYear('foo') on backward run container_id:roottestcleanupafterstart-gw6-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestcleanupafterstart-gw6-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Skipping toDayOfYear Checking bin('foo') Executing query select bin('foo') on backward Stdout:10 run container_id:roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select bin('foo') on upstream Stdout:10 OK bin Checking bitSlice('foo') Executing query select bitSlice('foo') on backward Skipping bitSlice Checking JSONExtract('foo') Executing query select JSONExtract('foo') on backward Skipping JSONExtract Checking tumbleStart('foo') Executing query select tumbleStart('foo') on backward Skipping tumbleStart Checking bitShiftRight('foo') Executing query select bitShiftRight('foo') on backward Skipping bitShiftRight Checking tan('foo') Executing query select tan('foo') on backward Skipping tan Checking tryBase58Decode('foo') Executing query select tryBase58Decode('foo') on backward Executing query select tryBase58Decode('foo') on upstream run container_id:roottestbackuprestoreoncluster-gw0-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestoreoncluster-gw0-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:10 run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:1240 Clickhouse process running. run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] OK tryBase58Decode Checking bitmapMax('foo') Executing query select bitmapMax('foo') on backward Stdout:1240 Executing query select 20 on node Stderr: Container roottestbuildsetsfrommultiplethreads-gw3-node-1 Stopping Stderr: Container roottestbuildsetsfrommultiplethreads-gw3-node-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_build_sets_from_multiple_threads/_instances-0-gw3/node/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_build_sets_from_multiple_threads/_instances-0-gw3/node/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_build_sets_from_multiple_threads/_instances-0-gw3/.env --project-name roottestbuildsetsfrommultiplethreads-gw3 --file /ClickHouse/tests/integration/test_build_sets_from_multiple_threads/_instances-0-gw3/node/docker-compose.yml down --volumes] Skipping bitmapMax Checking h3GetOriginIndexFromUnidirectionalEdge('foo') Executing query select h3GetOriginIndexFromUnidirectionalEdge('foo') on backward Skipping h3GetOriginIndexFromUnidirectionalEdge Checking ngramDistanceCaseInsensitiveUTF8('foo') Executing query select ngramDistanceCaseInsensitiveUTF8('foo') on backward run container_id:roottestcleanupafterstart-gw6-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestcleanupafterstart-gw6-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Skipping ngramDistanceCaseInsensitiveUTF8 Checking toDecimalString('foo') Executing query select toDecimalString('foo') on backward run container_id:roottestcleanupafterstart-gw6-node1-1 detach:False nothrow:False cmd: ['bash', '-c', 'mv /var/lib/clickhouse/data/default/test_table/20200101_0_0_0 /var/lib/clickhouse/data/default/test_table/delete_tmp_20200101_0_0_0'] Command:[docker exec --privileged roottestcleanupafterstart-gw6-node1-1 bash -c mv /var/lib/clickhouse/data/default/test_table/20200101_0_0_0 /var/lib/clickhouse/data/default/test_table/delete_tmp_20200101_0_0_0] Skipping toDecimalString Checking URLPathHierarchy('foo') Executing query select URLPathHierarchy('foo') on backward run container_id:roottestcleanupafterstart-gw6-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestcleanupafterstart-gw6-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] run container_id:roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] No clickhouse process running. Start new one. http://localhost:None "POST /v1.46/containers/roottestcleanupafterstart-gw6-node1-1/exec HTTP/1.1" 201 74 Executing query select URLPathHierarchy('foo') on upstream http://localhost:None "POST /v1.46/exec/58ca0c615b384cfa78ac5dd1ddfc1a905957a3e8af13f1426947865c4e01ece4/start HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/exec/58ca0c615b384cfa78ac5dd1ddfc1a905957a3e8af13f1426947865c4e01ece4/json HTTP/1.1" 200 586 Stdout:10 OK URLPathHierarchy Checking h3HexRing('foo') Executing query select h3HexRing('foo') on backward Skipping h3HexRing Checking h3GetUnidirectionalEdgesFromHexagon('foo') Executing query select h3GetUnidirectionalEdgesFromHexagon('foo') on backward Stderr: Container roottestbuildsetsfrommultiplethreads-gw3-node-1 Stopping Stderr: Container roottestbuildsetsfrommultiplethreads-gw3-node-1 Stopped Stderr: Container roottestbuildsetsfrommultiplethreads-gw3-node-1 Removing Stderr: Container roottestbuildsetsfrommultiplethreads-gw3-node-1 Removed Stderr: Network roottestbuildsetsfrommultiplethreads-gw3_default Removing Stderr: Network roottestbuildsetsfrommultiplethreads-gw3_default Removed Cleanup called Docker networks for project roottestbuildsetsfrommultiplethreads-gw3 are NETWORK ID NAME DRIVER SCOPE Skipping h3GetUnidirectionalEdgesFromHexagon Checking simpleJSONExtractUInt('foo') Executing query select simpleJSONExtractUInt('foo') on backward Docker containers for project roottestbuildsetsfrommultiplethreads-gw3 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Executing query select 20 on node Docker volumes for project roottestbuildsetsfrommultiplethreads-gw3 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestbuildsetsfrommultiplethreads-gw3-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestbuildsetsfrommultiplethreads-gw3 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Skipping simpleJSONExtractUInt Checking normalizedQueryHash('foo') Executing query select normalizedQueryHash('foo') on backward Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:1 Volumes pruned: 1 test_concurrent_queries_for_all_users_restriction/test.py::test_exception_message Running tests in /ClickHouse/tests/integration/test_concurrent_queries_for_all_users_restriction/test.py Cluster start called. is_up=False Docker networks for project roottestconcurrentqueriesforallusersrestriction-gw3 are NETWORK ID NAME DRIVER SCOPE Executing query select normalizedQueryHash('foo') on upstream Docker containers for project roottestconcurrentqueriesforallusersrestriction-gw3 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestconcurrentqueriesforallusersrestriction-gw3 are DRIVER VOLUME NAME Cleanup called Docker networks for project roottestconcurrentqueriesforallusersrestriction-gw3 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestconcurrentqueriesforallusersrestriction-gw3 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES OK normalizedQueryHash Checking bitTestAll('foo') Executing query select bitTestAll('foo') on backward Docker volumes for project roottestconcurrentqueriesforallusersrestriction-gw3 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestconcurrentqueriesforallusersrestriction-gw3-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestconcurrentqueriesforallusersrestriction-gw3 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Skipping bitTestAll Checking pi('foo') Executing query select pi('foo') on backward Stdout:1 Volumes pruned: 1 Setup directory for instance: node1 Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files [] to /ClickHouse/tests/integration/test_concurrent_queries_for_all_users_restriction/_instances-0-gw3/node1/configs/config.d Setup database dir /ClickHouse/tests/integration/test_concurrent_queries_for_all_users_restriction/_instances-0-gw3/node1/database Setup logs dir /ClickHouse/tests/integration/test_concurrent_queries_for_all_users_restriction/_instances-0-gw3/node1/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw'} stored in /ClickHouse/tests/integration/test_concurrent_queries_for_all_users_restriction/_instances-0-gw3/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_concurrent_queries_for_all_users_restriction/_instances-0-gw3/.env --project-name roottestconcurrentqueriesforallusersrestriction-gw3 --file /ClickHouse/tests/integration/test_concurrent_queries_for_all_users_restriction/_instances-0-gw3/node1/docker-compose.yml pull] Skipping pi Checking sleepEachRow('foo') Executing query select sleepEachRow('foo') on backward run container_id:roottestbackuprestoreoncluster-gw0-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestoreoncluster-gw0-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Skipping sleepEachRow Checking __bitSwapLastTwo('foo') Executing query select __bitSwapLastTwo('foo') on backward Stdout:10 Skipping __bitSwapLastTwo Checking fromModifiedJulianDay('foo') Executing query select fromModifiedJulianDay('foo') on backward Skipping fromModifiedJulianDay Checking CRC32('foo') Executing query select CRC32('foo') on backward Executing query select 20 on node Executing query select CRC32('foo') on upstream OK CRC32 Checking multiSearchFirstPositionUTF8('foo') Executing query select multiSearchFirstPositionUTF8('foo') on backward Skipping multiSearchFirstPositionUTF8 Checking addMilliseconds('foo') Executing query select addMilliseconds('foo') on backward Skipping addMilliseconds Checking byteSize('foo') Executing query select byteSize('foo') on backward run container_id:roottestcleanupafterstart-gw6-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestcleanupafterstart-gw6-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] run container_id:roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:773 Clickhouse process running. run container_id:roottestcleanupafterstart-gw6-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestcleanupafterstart-gw6-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:10 Stdout:773 Executing query select 20 on node1 run container_id:roottestcleanupafterstart-gw6-node1-1 detach:False nothrow:False cmd: ['bash', '-c', 'ls /var/lib/clickhouse/data/default/test_table/'] Command:[docker exec --privileged roottestcleanupafterstart-gw6-node1-1 bash -c ls /var/lib/clickhouse/data/default/test_table/] Executing query select byteSize('foo') on upstream Stdout:20200101_0_0_0 Stdout:detached Stdout:format_version.txt Executing query SELECT count() FROM test_table on node1 OK byteSize Checking intExp2('foo') Executing query select intExp2('foo') on backward Executing query select 20 on node Skipping intExp2 Checking assumeNotNull('foo') Executing query select assumeNotNull('foo') on backward run container_id:roottestcleanupafterstart-gw6-node1-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "Removing temporary directory" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true'] Command:[docker exec roottestcleanupafterstart-gw6-node1-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "Removing temporary directory" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true] Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.04.02 03:35:08.107730 [ 803 ] {} default.test_table (88752b86-ec59-4f3a-a581-dbd8f3ca6de6): Removing temporary directory /var/lib/clickhouse/store/887/88752b86-ec59-4f3a-a581-dbd8f3ca6de6/delete_tmp_20200101_0_0_0/ run container_id:roottestcleanupafterstart-gw6-node1-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "delete_tmp_20200101_0_0_0" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true'] Command:[docker exec roottestcleanupafterstart-gw6-node1-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "delete_tmp_20200101_0_0_0" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true] Executing query select assumeNotNull('foo') on upstream Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.04.02 03:35:08.107730 [ 803 ] {} default.test_table (88752b86-ec59-4f3a-a581-dbd8f3ca6de6): Removing temporary directory /var/lib/clickhouse/store/887/88752b86-ec59-4f3a-a581-dbd8f3ca6de6/delete_tmp_20200101_0_0_0/ Command:[docker compose --env-file /ClickHouse/tests/integration/test_cleanup_after_start/_instances-0-gw6/.env --project-name roottestcleanupafterstart-gw6 --file /ClickHouse/tests/integration/test_cleanup_after_start/_instances-0-gw6/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml stop --timeout 20] [gw6] PASSED test_cleanup_after_start/test.py::test_old_dirs_cleanup OK assumeNotNull Checking wordShingleMinHashArg('foo') Executing query select wordShingleMinHashArg('foo') on backward Executing query select wordShingleMinHashArg('foo') on upstream run container_id:roottestbackuprestoreoncluster-gw0-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestoreoncluster-gw0-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] OK wordShingleMinHashArg Checking ngramMinHashArgUTF8('foo') Executing query select ngramMinHashArgUTF8('foo') on backward Executing query SELECT status FROM system.backups WHERE id='cc85aefc-6a4e-45a6-9133-64badb5ea938' AND status == 'CREATING_BACKUP' on node1 Executing query select 20 on node Executing query SELECT status FROM system.backups WHERE id='cc85aefc-6a4e-45a6-9133-64badb5ea938' on node1 Executing query select ngramMinHashArgUTF8('foo') on upstream OK ngramMinHashArgUTF8 Checking ngramMinHashArgCaseInsensitive('foo') Executing query select ngramMinHashArgCaseInsensitive('foo') on backward run container_id:roottestbackuprestoreoncluster-gw0-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestoreoncluster-gw0-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] run container_id:roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select ngramMinHashArgCaseInsensitive('foo') on upstream No clickhouse process running. Start new one. http://localhost:None "POST /v1.46/containers/roottestbackuprestoreoncluster-gw0-node2-1/exec HTTP/1.1" 201 74 Stdout:10 http://localhost:None "POST /v1.46/exec/f56794e17417c4ab7a7855226b2a33d168eb63476b6ebd784bddc255793ad920/start HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/exec/f56794e17417c4ab7a7855226b2a33d168eb63476b6ebd784bddc255793ad920/json HTTP/1.1" 200 586 OK ngramMinHashArgCaseInsensitive Checking cutToFirstSignificantSubdomainRFC('foo') Executing query select cutToFirstSignificantSubdomainRFC('foo') on backward Executing query select cutToFirstSignificantSubdomainRFC('foo') on upstream OK cutToFirstSignificantSubdomainRFC Checking wordShingleMinHashUTF8('foo') Executing query select wordShingleMinHashUTF8('foo') on backward Executing query select 20 on node Executing query select wordShingleMinHashUTF8('foo') on upstream OK wordShingleMinHashUTF8 Checking arrayDifference('foo') Executing query select arrayDifference('foo') on backward Skipping arrayDifference Checking wordShingleMinHash('foo') Executing query select wordShingleMinHash('foo') on backward Executing query select wordShingleMinHash('foo') on upstream OK wordShingleMinHash Checking ngramMinHashCaseInsensitiveUTF8('foo') Executing query select ngramMinHashCaseInsensitiveUTF8('foo') on backward Executing query select 20 on node Executing query select ngramMinHashCaseInsensitiveUTF8('foo') on upstream Stderr: Container roottestcleanupafterstart-gw6-node1-1 Stopping Stderr: Container roottestcleanupafterstart-gw6-node1-1 Stopped Stderr: Container roottestcleanupafterstart-gw6-zoo3-1 Stopping Stderr: Container roottestcleanupafterstart-gw6-zoo1-1 Stopping Stderr: Container roottestcleanupafterstart-gw6-zoo2-1 Stopping Stderr: Container roottestcleanupafterstart-gw6-zoo2-1 Stopped Stderr: Container roottestcleanupafterstart-gw6-zoo1-1 Stopped Stderr: Container roottestcleanupafterstart-gw6-zoo3-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_cleanup_after_start/_instances-0-gw6/node1/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_cleanup_after_start/_instances-0-gw6/node1/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_cleanup_after_start/_instances-0-gw6/.env --project-name roottestcleanupafterstart-gw6 --file /ClickHouse/tests/integration/test_cleanup_after_start/_instances-0-gw6/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml down --volumes] OK ngramMinHashCaseInsensitiveUTF8 Checking toMonth('foo') Executing query select toMonth('foo') on backward Skipping toMonth Checking ngramMinHash('foo') Executing query select ngramMinHash('foo') on backward run container_id:roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] run container_id:roottestbackuprestoreoncluster-gw0-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestoreoncluster-gw0-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select ngramMinHash('foo') on upstream Stdout:10 Stdout:883 Clickhouse process running. run container_id:roottestbackuprestoreoncluster-gw0-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestoreoncluster-gw0-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:883 Executing query select 20 on node2 OK ngramMinHash Checking formatRowNoNewline('foo') Executing query select formatRowNoNewline('foo') on backward Executing query DROP TABLE tbl ON CLUSTER 'cluster' SYNC on node1 Skipping formatRowNoNewline Checking wordShingleSimHashCaseInsensitiveUTF8('foo') Executing query select wordShingleSimHashCaseInsensitiveUTF8('foo') on backward Executing query select wordShingleSimHashCaseInsensitiveUTF8('foo') on upstream OK wordShingleSimHashCaseInsensitiveUTF8 Checking wordShingleSimHashUTF8('foo') Executing query select wordShingleSimHashUTF8('foo') on backward Executing query RESTORE TABLE tbl ON CLUSTER 'cluster' FROM Disk('backups', '26') on node1 Executing query select 20 on node Executing query select wordShingleSimHashUTF8('foo') on upstream OK wordShingleSimHashUTF8 Checking arrayCumSum('foo') Executing query select arrayCumSum('foo') on backward Stderr: Container roottestcleanupafterstart-gw6-node1-1 Stopping Stderr: Container roottestcleanupafterstart-gw6-node1-1 Stopped Stderr: Container roottestcleanupafterstart-gw6-node1-1 Removing Stderr: Container roottestcleanupafterstart-gw6-node1-1 Removed Stderr: Container roottestcleanupafterstart-gw6-zoo1-1 Stopping Stderr: Container roottestcleanupafterstart-gw6-zoo3-1 Stopping Stderr: Container roottestcleanupafterstart-gw6-zoo2-1 Stopping Stderr: Container roottestcleanupafterstart-gw6-zoo1-1 Stopped Stderr: Container roottestcleanupafterstart-gw6-zoo1-1 Removing Stderr: Container roottestcleanupafterstart-gw6-zoo2-1 Stopped Stderr: Container roottestcleanupafterstart-gw6-zoo2-1 Removing Stderr: Container roottestcleanupafterstart-gw6-zoo3-1 Stopped Stderr: Container roottestcleanupafterstart-gw6-zoo3-1 Removing Stderr: Container roottestcleanupafterstart-gw6-zoo1-1 Removed Stderr: Container roottestcleanupafterstart-gw6-zoo2-1 Removed Stderr: Container roottestcleanupafterstart-gw6-zoo3-1 Removed Stderr: Network roottestcleanupafterstart-gw6_default Removing Stderr: Network roottestcleanupafterstart-gw6_default Removed Cleanup called Executing query SYSTEM SYNC REPLICA tbl on node1 Docker networks for project roottestcleanupafterstart-gw6 are NETWORK ID NAME DRIVER SCOPE Skipping arrayCumSum Checking wordShingleSimHash('foo') Executing query select wordShingleSimHash('foo') on backward Docker containers for project roottestcleanupafterstart-gw6 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestcleanupafterstart-gw6 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestcleanupafterstart-gw6-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestcleanupafterstart-gw6 Trying to prune unused networks... Executing query SELECT * FROM tbl ORDER BY x on node1 Trying to prune unused images... Command:[docker image prune -f] Executing query select wordShingleSimHash('foo') on upstream Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:1 Volumes pruned: 1 Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw0] PASSED test_backup_restore_on_cluster/test.py::test_shutdown_waits_for_backup OK wordShingleSimHash Checking finalizeAggregation('foo') Executing query select finalizeAggregation('foo') on backward Skipping finalizeAggregation Checking ngramSimHashCaseInsensitiveUTF8('foo') Executing query select ngramSimHashCaseInsensitiveUTF8('foo') on backward Executing query select ngramSimHashCaseInsensitiveUTF8('foo') on upstream Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 OK ngramSimHashCaseInsensitiveUTF8 Checking ngramSimHashCaseInsensitive('foo') Executing query select ngramSimHashCaseInsensitive('foo') on backward Executing query select 20 on node Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Executing query select ngramSimHashCaseInsensitive('foo') on upstream OK ngramSimHashCaseInsensitive Checking ngramSimHash('foo') Executing query select ngramSimHash('foo') on backward Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select ngramSimHash('foo') on upstream run container_id:roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:10 OK ngramSimHash Checking toTimezone('foo') Executing query select toTimezone('foo') on backward Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Stderr: node1 Pulling Stderr: node1 Pulled ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_concurrent_queries_for_all_users_restriction/_instances-0-gw3/.env --project-name roottestconcurrentqueriesforallusersrestriction-gw3 --file /ClickHouse/tests/integration/test_concurrent_queries_for_all_users_restriction/_instances-0-gw3/node1/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_concurrent_queries_for_all_users_restriction/_instances-0-gw3/.env --project-name roottestconcurrentqueriesforallusersrestriction-gw3 --file /ClickHouse/tests/integration/test_concurrent_queries_for_all_users_restriction/_instances-0-gw3/node1/docker-compose.yml up -d --no-recreate] Skipping toTimezone Checking currentDatabase('foo') Executing query select currentDatabase('foo') on backward Skipping currentDatabase Checking splitByRegexp('foo') Executing query select splitByRegexp('foo') on backward test_backup_restore_on_cluster/test.py::test_system_functions Executing query CREATE FUNCTION linear_equation AS (x, k, b) -> k*x + b; on node1 Skipping splitByRegexp Checking rowNumberInBlock('foo') Executing query select rowNumberInBlock('foo') on backward Executing query CREATE FUNCTION parity_str AS (n) -> if(n % 2, 'odd', 'even'); on node1 Skipping rowNumberInBlock Checking splitByString('foo') Executing query select splitByString('foo') on backward Executing query BACKUP TABLE system.functions ON CLUSTER 'cluster' TO Disk('backups', '27') on node1 Skipping splitByString Checking s2CapContains('foo') Executing query select s2CapContains('foo') on backward Executing query select 20 on node Skipping s2CapContains Checking identity('foo') Executing query select identity('foo') on backward Executing query select identity('foo') on upstream Executing query DROP FUNCTION linear_equation on node1 Stderr: Network roottestconcurrentqueriesforallusersrestriction-gw3_default Creating Stderr: Network roottestconcurrentqueriesforallusersrestriction-gw3_default Created Stderr: Container roottestconcurrentqueriesforallusersrestriction-gw3-node1-1 Creating Stderr: Container roottestconcurrentqueriesforallusersrestriction-gw3-node1-1 Created Stderr: Container roottestconcurrentqueriesforallusersrestriction-gw3-node1-1 Starting Stderr: Container roottestconcurrentqueriesforallusersrestriction-gw3-node1-1 Started ClickHouse instance created get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottestconcurrentqueriesforallusersrestriction-gw3-node1-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottestconcurrentqueriesforallusersrestriction-gw3-node1-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node1, ip: 172.16.6.2... http://localhost:None "GET /v1.46/containers/roottestconcurrentqueriesforallusersrestriction-gw3-node1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/d08b4b98881b75bd2657c1e41ad51248ff7e2fb7f56cd7f5e958b0cfbe9d3f62/json HTTP/1.1" 200 None Executing query DROP FUNCTION parity_str on node1 OK identity Checking ascii('foo') Executing query select ascii('foo') on backward Executing query SELECT name FROM system.functions WHERE name='parity_str' on node2 Executing query select ascii('foo') on upstream http://localhost:None "GET /v1.46/containers/d08b4b98881b75bd2657c1e41ad51248ff7e2fb7f56cd7f5e958b0cfbe9d3f62/json HTTP/1.1" 200 None Stderr: zoo2 Skipped - Image is already being pulled by node2 Stderr: zoo3 Skipped - Image is already being pulled by node2 Stderr: zoo1 Skipped - Image is already being pulled by node2 Stderr: node2 Pulling Stderr: node1 Pulling Stderr: node2 Pulled Stderr: node1 Pulled Setup ZooKeeper Creating internal ZooKeeper dirs: ['/ClickHouse/tests/integration/test_backward_compatibility/_instances-vertical_merges_from_compact_parts-0-gw1/keeper1/log', '/ClickHouse/tests/integration/test_backward_compatibility/_instances-vertical_merges_from_compact_parts-0-gw1/keeper1/config', '/ClickHouse/tests/integration/test_backward_compatibility/_instances-vertical_merges_from_compact_parts-0-gw1/keeper1/coordination', '/ClickHouse/tests/integration/test_backward_compatibility/_instances-vertical_merges_from_compact_parts-0-gw1/keeper2/log', '/ClickHouse/tests/integration/test_backward_compatibility/_instances-vertical_merges_from_compact_parts-0-gw1/keeper2/config', '/ClickHouse/tests/integration/test_backward_compatibility/_instances-vertical_merges_from_compact_parts-0-gw1/keeper2/coordination', '/ClickHouse/tests/integration/test_backward_compatibility/_instances-vertical_merges_from_compact_parts-0-gw1/keeper3/log', '/ClickHouse/tests/integration/test_backward_compatibility/_instances-vertical_merges_from_compact_parts-0-gw1/keeper3/config', '/ClickHouse/tests/integration/test_backward_compatibility/_instances-vertical_merges_from_compact_parts-0-gw1/keeper3/coordination'] Command:[docker compose --project-name roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1 --env-file /ClickHouse/tests/integration/test_backward_compatibility/_instances-vertical_merges_from_compact_parts-0-gw1/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --verbose up -d] Executing query RESTORE TABLE system.functions ON CLUSTER 'cluster' FROM Disk('backups', '27') on node1 OK ascii Checking s2RectUnion('foo') Executing query select s2RectUnion('foo') on backward http://localhost:None "GET /v1.46/containers/d08b4b98881b75bd2657c1e41ad51248ff7e2fb7f56cd7f5e958b0cfbe9d3f62/json HTTP/1.1" 200 None Skipping s2RectUnion Checking ngramDistanceUTF8('foo') Executing query select ngramDistanceUTF8('foo') on backward Skipping ngramDistanceUTF8 Checking MD4('foo') Executing query select MD4('foo') on backward Executing query SELECT number, linear_equation(number, 2, 1) FROM numbers(3) on node1 http://localhost:None "GET /v1.46/containers/d08b4b98881b75bd2657c1e41ad51248ff7e2fb7f56cd7f5e958b0cfbe9d3f62/json HTTP/1.1" 200 None run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select MD4('foo') on upstream Executing query SELECT number, parity_str(number) FROM numbers(3) on node1 Current start attempt failed. Will kill 1240 just in case. run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:True cmd: ['bash', '-c', 'kill -9 1240'] Command:[docker exec -u root roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c kill -9 1240] OK MD4 Checking moduloOrZero('foo') Executing query select moduloOrZero('foo') on backward http://localhost:None "GET /v1.46/containers/d08b4b98881b75bd2657c1e41ad51248ff7e2fb7f56cd7f5e958b0cfbe9d3f62/json HTTP/1.1" 200 None ClickHouse node1 started Executing query create table nums (number UInt64) ENGINE = MergeTree() order by tuple() on node1 Stderr:bash: line 0: kill: (1240) - No such process Exitcode:1 Executing query SELECT number, linear_equation(number, 2, 1) FROM numbers(3) on node2 run container_id:roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Skipping moduloOrZero Checking sipHash64Keyed('foo') Executing query select sipHash64Keyed('foo') on backward Stdout:10 Executing query insert into nums values (0), (1) on node1 Executing query SELECT number, parity_str(number) FROM numbers(3) on node2 Skipping sipHash64Keyed Checking extractAll('foo') Executing query select extractAll('foo') on backward Executing query select number from nums order by number on node1 Executing query SELECT name FROM system.functions WHERE name='parity_str' on node2 Stderr:time="2025-04-02T03:35:12Z" level=trace msg="Docker Desktop integration not enabled" Stderr: Network roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1_default Creating Stderr: Network roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1_default Created Stderr: Container roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-zoo3-1 Creating Stderr: Container roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-zoo1-1 Creating Stderr: Container roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-zoo2-1 Creating Stderr: Container roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-zoo3-1 Created Stderr: Container roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-zoo1-1 Created Stderr: Container roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-zoo2-1 Created Stderr: Container roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-zoo3-1 Starting Stderr: Container roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-zoo1-1 Starting Stderr: Container roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-zoo2-1 Starting Stderr: Container roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-zoo3-1 Started Stderr: Container roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-zoo2-1 Started Stderr: Container roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-zoo1-1 Started Stderr:time="2025-04-02T03:35:13Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T03:35:13Z" level=debug msg="otel error" error="" Wait ZooKeeper to start get_instance_ip instance_name=zoo1 Skipping extractAll http://localhost:None "GET /v1.46/containers/roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-zoo1-1/json HTTP/1.1" 200 None Checking age('foo') Executing query select age('foo') on backward get_kazoo_client: zoo1, ip:172.16.7.2, port:2181, use_ssl:False Connecting to 172.16.7.2(172.16.7.2):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Executing query select sleep(2) on node1 Executing query select sleep(2) on node1 Executing query select sleep(2) on node1 Executing query SELECT number, parity_str(number) FROM numbers(3) on node2 Skipping age Checking dictGetOrNull('foo') Executing query select dictGetOrNull('foo') on backward Executing query select sleep(2) on node1 Executing query DROP FUNCTION linear_equation on node1 Connecting to 172.16.7.2(172.16.7.2):2181, use_ssl: False Connection dropped: socket connection error: Connection refused run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Skipping dictGetOrNull Checking regionToDistrict('foo') Executing query select regionToDistrict('foo') on backward Executing query select sleep(2) on node1 Executing query DROP FUNCTION parity_str on node1 No clickhouse process running. Start new one. http://localhost:None "POST /v1.46/containers/roottestbackwardcompatibilityconvertordinary-gw5-node-1/exec HTTP/1.1" 201 74 http://localhost:None "POST /v1.46/exec/a460fa989086f64bd262d3dcec9fdc09387925422f89ebe8af56538874b77b87/start HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/exec/a460fa989086f64bd262d3dcec9fdc09387925422f89ebe8af56538874b77b87/json HTTP/1.1" 200 586 Skipping regionToDistrict Checking unbin('foo') Executing query select unbin('foo') on backward Executing query select sleep(2) on node1 Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw0] PASSED test_backup_restore_on_cluster/test.py::test_system_functions Executing query select unbin('foo') on upstream Executing query select sleep(2) on node1 Connecting to 172.16.7.2(172.16.7.2):2181, use_ssl: False Connection dropped: socket connection error: Connection refused OK unbin Checking reinterpret('foo') Executing query select reinterpret('foo') on backward Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select sleep(2) on node1 Skipping reinterpret Checking modulo('foo') Executing query select modulo('foo') on backward Executing query select sleep(2) on node1 Skipping modulo Checking lowCardinalityKeys('foo') Executing query select lowCardinalityKeys('foo') on backward Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Executing query select sleep(2) on node1 Skipping lowCardinalityKeys Checking nullInIgnoreSet('foo') Executing query select nullInIgnoreSet('foo') on backward Connecting to 172.16.7.2(172.16.7.2):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Executing query select sleep(2) on node1 Skipping nullInIgnoreSet Checking geoToS2('foo') Executing query select geoToS2('foo') on backward Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select sleep(2) on node1 Skipping geoToS2 Checking blockSerializedSize('foo') Executing query select blockSerializedSize('foo') on backward run container_id:roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select blockSerializedSize('foo') on upstream Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Stdout:10 OK blockSerializedSize Checking dateTrunc('foo') Executing query select dateTrunc('foo') on backward Connecting to 172.16.7.2(172.16.7.2):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Skipping dateTrunc Checking h3ExactEdgeLengthRads('foo') Executing query select h3ExactEdgeLengthRads('foo') on backward test_backup_restore_on_cluster/test.py::test_system_users Executing query CREATE USER u1 SETTINGS custom_a=123 on node1 Skipping h3ExactEdgeLengthRads Checking JSONExtractBool('foo') Executing query select JSONExtractBool('foo') on backward Executing query GRANT SELECT ON tbl TO u1 on node1 Executing query select JSONExtractBool('foo') on upstream Executing query CREATE USER u2 SETTINGS allow_backup=false on node1 OK JSONExtractBool Checking tupleHammingDistance('foo') Executing query select tupleHammingDistance('foo') on backward Executing query select number from remote('node1', 'default', 'nums') on node1 Executing query GRANT CLUSTER ON *.* TO u2 on node1 Skipping tupleHammingDistance Checking ngramSearchUTF8('foo') Executing query select ngramSearchUTF8('foo') on backward Executing query select number from remote('node1', 'default', 'nums') on node1 Executing query BACKUP TABLE system.users ON CLUSTER 'cluster' TO Disk('backups', '28') on node1 Skipping ngramSearchUTF8 Checking minSampleSizeConversion('foo') Executing query select minSampleSizeConversion('foo') on backward Connecting to 172.16.7.2(172.16.7.2):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Skipping minSampleSizeConversion Checking jumpConsistentHash('foo') Executing query select jumpConsistentHash('foo') on backward run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select number from remote('node1', 'default', 'nums') on node1 Executing query GRANT BACKUP ON system.users TO u2 on node1 Skipping jumpConsistentHash Checking parseTimeDelta('foo') Executing query select parseTimeDelta('foo') on backward Stdout:2020 Clickhouse process running. run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select number from remote('node1', 'default', 'nums') on node1 Executing query BACKUP TABLE system.users ON CLUSTER 'cluster' TO Disk('backups', '28') on node1 Stdout:2020 Executing query select 20 on node Skipping parseTimeDelta Checking formatReadableSize('foo') Executing query select formatReadableSize('foo') on backward Command:[docker compose --env-file /ClickHouse/tests/integration/test_concurrent_queries_for_all_users_restriction/_instances-0-gw3/.env --project-name roottestconcurrentqueriesforallusersrestriction-gw3 --file /ClickHouse/tests/integration/test_concurrent_queries_for_all_users_restriction/_instances-0-gw3/node1/docker-compose.yml stop --timeout 20] [gw3] PASSED test_concurrent_queries_for_all_users_restriction/test.py::test_exception_message Skipping formatReadableSize Checking toNullable('foo') Executing query select toNullable('foo') on backward Executing query select toNullable('foo') on upstream Executing query DROP USER u1 on node1 OK toNullable Checking concat('foo') Executing query select concat('foo') on backward Executing query RESTORE TABLE system.users ON CLUSTER 'cluster' FROM Disk('backups', '28') on node1 Skipping concat Checking demangle('foo') Executing query select demangle('foo') on backward Executing query GRANT CREATE USER ON *.* TO u2 on node1 Skipping demangle Checking arrayStringConcat('foo') Executing query select arrayStringConcat('foo') on backward Executing query RESTORE TABLE system.users ON CLUSTER 'cluster' FROM Disk('backups', '28') on node1 Skipping arrayStringConcat Checking splitByNonAlpha('foo') Executing query select splitByNonAlpha('foo') on backward run container_id:roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query GRANT SELECT ON tbl TO u2 WITH GRANT OPTION on node1 Executing query select splitByNonAlpha('foo') on upstream Stdout:10 Executing query select 20 on node Executing query RESTORE TABLE system.users ON CLUSTER 'cluster' FROM Disk('backups', '28') on node1 OK splitByNonAlpha Checking hasSubstr('foo') Executing query select hasSubstr('foo') on backward Connecting to 172.16.7.2(172.16.7.2):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Skipping hasSubstr Checking toUInt64OrDefault('foo') Executing query select toUInt64OrDefault('foo') on backward Executing query select toUInt64OrDefault('foo') on upstream Executing query SHOW CREATE USER u1 on node1 OK toUInt64OrDefault Checking __bitBoolMaskAnd('foo') Executing query select __bitBoolMaskAnd('foo') on backward Executing query select sleep(2) on node1 Executing query select sleep(2) on node1 Skipping __bitBoolMaskAnd Checking h3GetIndexesFromUnidirectionalEdge('foo') Executing query select h3GetIndexesFromUnidirectionalEdge('foo') on backward Executing query SHOW GRANTS FOR u1 on node1 Executing query select sleep(2) on node1 Executing query select sleep(2) on node1 Skipping h3GetIndexesFromUnidirectionalEdge Checking sipHash64('foo') Executing query select sipHash64('foo') on backward Executing query select sleep(2) on node1 Executing query select sleep(2) on node1 Executing query select sleep(2) on node1 Executing query select sleep(2) on node1 Executing query select sipHash64('foo') on upstream Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw0] PASSED test_backup_restore_on_cluster/test.py::test_system_users Executing query select sleep(2) on node1 Executing query select sleep(2) on node1 Executing query select sleep(2) on node1 Executing query select sleep(2) on node1 Executing query select sleep(2) on node1 OK sipHash64 Checking randNormal('foo') Executing query select randNormal('foo') on backward Executing query select sleep(2) on node1 Executing query select sleep(2) on node1 Executing query select 20 on node Executing query select sleep(2) on node1 Executing query select sleep(2) on node1 Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Skipping randNormal Checking catboostEvaluate('foo') Executing query select catboostEvaluate('foo') on backward Executing query select sleep(2) on node1 Skipping catboostEvaluate Checking divide('foo') Executing query select divide('foo') on backward Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Skipping divide Checking encrypt('foo') Executing query select encrypt('foo') on backward Skipping encrypt Checking intDivOrZero('foo') Executing query select intDivOrZero('foo') on backward Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Skipping intDivOrZero Checking minus('foo') Executing query select minus('foo') on backward Skipping minus Checking ngramMinHashUTF8('foo') Executing query select ngramMinHashUTF8('foo') on backward run container_id:roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Executing query select ngramMinHashUTF8('foo') on upstream run container_id:roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] OK ngramMinHashUTF8 Checking sqrt('foo') Executing query select sqrt('foo') on backward No clickhouse process running. Start new one. http://localhost:None "POST /v1.46/containers/roottestbackuprestorenewshutdownwaitbackup-gw2-node-1/exec HTTP/1.1" 201 74 Executing query select 20 on node http://localhost:None "POST /v1.46/exec/04ee91ffdd50596e1e3eb042d38dd6ad8654e8d01b5ee0971ecb39b773438435/start HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/exec/04ee91ffdd50596e1e3eb042d38dd6ad8654e8d01b5ee0971ecb39b773438435/json HTTP/1.1" 200 586 test_backup_restore_on_cluster/test.py::test_table_in_replicated_database_with_not_synced_def Executing query CREATE DATABASE mydb ON CLUSTER 'cluster' ENGINE=Replicated('/clickhouse/path/','{shard}','{replica}') on node1 Skipping sqrt Checking ngramMinHashArgCaseInsensitiveUTF8('foo') Executing query select ngramMinHashArgCaseInsensitiveUTF8('foo') on backward Executing query select ngramMinHashArgCaseInsensitiveUTF8('foo') on upstream OK ngramMinHashArgCaseInsensitiveUTF8 Checking endsWith('foo') Executing query select endsWith('foo') on backward Executing query CREATE TABLE mydb.tbl (x UInt8, y String) ENGINE=ReplicatedMergeTree ORDER BY tuple() on node1 Skipping endsWith Checking evalMLMethod('foo') Executing query select evalMLMethod('foo') on backward Skipping evalMLMethod Checking emptyArrayInt32('foo') Executing query select emptyArrayInt32('foo') on backward Skipping emptyArrayInt32 Checking extract('foo') Executing query select extract('foo') on backward Executing query ALTER TABLE mydb.tbl MODIFY COLUMN x String on node1 Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost Skipping extract Checking multiSearchFirstIndexCaseInsensitive('foo') Executing query select multiSearchFirstIndexCaseInsensitive('foo') on backward Connection dropped: outstanding heartbeat ping not received Transition to CONNECTING Zookeeper connection lost Skipping multiSearchFirstIndexCaseInsensitive Checking rowNumberInAllBlocks('foo') Executing query select rowNumberInAllBlocks('foo') on backward Executing query BACKUP DATABASE mydb ON CLUSTER 'cluster' TO Disk('backups', '29') on node2 Connecting to 172.16.2.3(172.16.2.3):2181, use_ssl: False Connecting to 172.16.2.3(172.16.2.3):2181, use_ssl: False Executing query select 20 on node Connecting to 172.16.7.2(172.16.7.2):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost Skipping rowNumberInAllBlocks Checking countDigits('foo') Executing query select countDigits('foo') on backward Skipping countDigits Checking replaceRegexpOne('foo') Executing query select replaceRegexpOne('foo') on backward Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED get_instance_ip instance_name=zoo2 http://localhost:None "GET /v1.46/containers/roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-zoo2-1/json HTTP/1.1" 200 None get_kazoo_client: zoo2, ip:172.16.7.4, port:2181, use_ssl:False Connecting to 172.16.7.4(172.16.7.4):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost Skipping replaceRegexpOne Checking factorial('foo') Executing query select factorial('foo') on backward Executing query DROP DATABASE mydb ON CLUSTER 'cluster' SYNC on node1 Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED get_instance_ip instance_name=zoo3 http://localhost:None "GET /v1.46/containers/roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-zoo3-1/json HTTP/1.1" 200 None get_kazoo_client: zoo3, ip:172.16.7.3, port:2181, use_ssl:False Connecting to 172.16.7.3(172.16.7.3):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost Skipping factorial Checking mapFilter('foo') Executing query select mapFilter('foo') on backward Skipping mapFilter Checking cbrt('foo') Executing query select cbrt('foo') on backward Skipping cbrt Checking toStartOfMinute('foo') Executing query select toStartOfMinute('foo') on backward Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED All instances of ZooKeeper started: ('zoo1', 'zoo2', 'zoo3') ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_backward_compatibility/_instances-vertical_merges_from_compact_parts-0-gw1/.env --project-name roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1 --file /ClickHouse/tests/integration/test_backward_compatibility/_instances-vertical_merges_from_compact_parts-0-gw1/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_backward_compatibility/_instances-vertical_merges_from_compact_parts-0-gw1/node2/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_backward_compatibility/_instances-vertical_merges_from_compact_parts-0-gw1/.env --project-name roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1 --file /ClickHouse/tests/integration/test_backward_compatibility/_instances-vertical_merges_from_compact_parts-0-gw1/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_backward_compatibility/_instances-vertical_merges_from_compact_parts-0-gw1/node2/docker-compose.yml up -d --no-recreate] Skipping toStartOfMinute Checking LpNormalize('foo') Executing query select LpNormalize('foo') on backward Executing query RESTORE DATABASE mydb ON CLUSTER 'cluster' FROM Disk('backups', '29') SETTINGS replica_num_in_backup=1 on node1 Skipping LpNormalize Checking tokens('foo') Executing query select tokens('foo') on backward run container_id:roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select tokens('foo') on upstream Stdout:1063 Clickhouse process running. run container_id:roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select 20 on node Stdout:1063 Executing query select 20 on node OK tokens Checking domain('foo') Executing query select domain('foo') on backward Executing query SELECT status FROM system.backup_log WHERE id='0d4abe713f324bc6b013778cacfec174' ORDER BY status on node Executing query select domain('foo') on upstream Executing query SELECT name, type FROM system.columns WHERE database='mydb' AND table='tbl' on node1 Stderr: Container roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-zoo2-1 Running Stderr: Container roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-zoo1-1 Running Stderr: Container roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-zoo3-1 Running Stderr: Container roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node2-1 Creating Stderr: Container roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node1-1 Creating Stderr: Container roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node1-1 Created Stderr: Container roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node2-1 Created Stderr: Container roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node1-1 Starting Stderr: Container roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node2-1 Starting Stderr: Container roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node1-1 Started Stderr: Container roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node2-1 Started ClickHouse instance created get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node1-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node1-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node1, ip: 172.16.7.5... http://localhost:None "GET /v1.46/containers/roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/20597af86768b8fdf37c13ebaed6aa6e672f100b9ffe8e4293a43db89cadb74e/json HTTP/1.1" 200 None OK domain Checking toDecimal64OrDefault('foo') Executing query select toDecimal64OrDefault('foo') on backward Executing query SELECT name, type FROM system.columns WHERE database='mydb' AND table='tbl' on node2 Executing query DROP TABLE tbl SYNC on node Skipping toDecimal64OrDefault Checking ngrams('foo') Executing query select ngrams('foo') on backward http://localhost:None "GET /v1.46/containers/20597af86768b8fdf37c13ebaed6aa6e672f100b9ffe8e4293a43db89cadb74e/json HTTP/1.1" 200 None Executing query DROP DATABASE mydb ON CLUSTER 'cluster' SYNC on node1 Executing query RESTORE TABLE tbl FROM Disk('backups', '0d4abe713f324bc6b013778cacfec174') on node Skipping ngrams Checking parseDateTime32BestEffortOrNull('foo') Executing query select parseDateTime32BestEffortOrNull('foo') on backward http://localhost:None "GET /v1.46/containers/20597af86768b8fdf37c13ebaed6aa6e672f100b9ffe8e4293a43db89cadb74e/json HTTP/1.1" 200 None Executing query select parseDateTime32BestEffortOrNull('foo') on upstream Executing query RESTORE DATABASE mydb ON CLUSTER 'cluster' FROM Disk('backups', '29') SETTINGS replica_num_in_backup=2 on node2 Executing query select 20 on node http://localhost:None "GET /v1.46/containers/20597af86768b8fdf37c13ebaed6aa6e672f100b9ffe8e4293a43db89cadb74e/json HTTP/1.1" 200 None ClickHouse node1 started get_instance_ip instance_name=node2 http://localhost:None "GET /v1.46/containers/roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node2-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node2 http://localhost:None "GET /v1.46/containers/roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node2-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node2, ip: 172.16.7.6... http://localhost:None "GET /v1.46/containers/roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node2-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b71deb33b62c30d2e070ba7aa8b3debd0c5c8df954a8b8a8a24dc17ace7b2b87/json HTTP/1.1" 200 None OK parseDateTime32BestEffortOrNull Checking cutToFirstSignificantSubdomainWithWWW('foo') Executing query select cutToFirstSignificantSubdomainWithWWW('foo') on backward Executing query select cutToFirstSignificantSubdomainWithWWW('foo') on upstream http://localhost:None "GET /v1.46/containers/b71deb33b62c30d2e070ba7aa8b3debd0c5c8df954a8b8a8a24dc17ace7b2b87/json HTTP/1.1" 200 None ClickHouse node2 started Executing query CREATE TABLE t_vertical_merges (id UInt64, v1 UInt64, v2 UInt64) ENGINE = ReplicatedMergeTree('/clickhouse/tables/0/t_vertical_merges', '0') ORDER BY id SETTINGS index_granularity = 50, vertical_merge_algorithm_min_rows_to_activate = 1, vertical_merge_algorithm_min_columns_to_activate = 1, min_bytes_for_wide_part = 0, min_rows_for_wide_part = 100 on node1 OK cutToFirstSignificantSubdomainWithWWW Checking wordShingleSimHashCaseInsensitive('foo') Executing query select wordShingleSimHashCaseInsensitive('foo') on backward Executing query CREATE TABLE t_vertical_merges (id UInt64, v1 UInt64, v2 UInt64) ENGINE = ReplicatedMergeTree('/clickhouse/tables/0/t_vertical_merges', '1') ORDER BY id SETTINGS index_granularity = 50, vertical_merge_algorithm_min_rows_to_activate = 1, vertical_merge_algorithm_min_columns_to_activate = 1, min_bytes_for_wide_part = 0, min_rows_for_wide_part = 100 on node2 Executing query select wordShingleSimHashCaseInsensitive('foo') on upstream Executing query SELECT name, type FROM system.columns WHERE database='mydb' AND table='tbl' on node1 Executing query INSERT INTO t_vertical_merges SELECT number, number, number FROM numbers(60) on node2 OK wordShingleSimHashCaseInsensitive Checking bitXor('foo') Executing query select bitXor('foo') on backward Executing query INSERT INTO t_vertical_merges SELECT number * 2, number, number FROM numbers(60) on node2 Executing query SELECT name, type FROM system.columns WHERE database='mydb' AND table='tbl' on node2 Executing query OPTIMIZE TABLE t_vertical_merges FINAL on node2 Skipping bitXor Checking lcm('foo') Executing query select lcm('foo') on backward Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw0] PASSED test_backup_restore_on_cluster/test.py::test_table_in_replicated_database_with_not_synced_def Executing query SYSTEM SYNC REPLICA t_vertical_merges on node1 Executing query select 20 on node Executing query SYSTEM FLUSH LOGS on node2 Skipping lcm Checking detectLanguage('foo') Executing query select detectLanguage('foo') on backward Stderr: Container roottestconcurrentqueriesforallusersrestriction-gw3-node1-1 Stopping Stderr: Container roottestconcurrentqueriesforallusersrestriction-gw3-node1-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_concurrent_queries_for_all_users_restriction/_instances-0-gw3/node1/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_concurrent_queries_for_all_users_restriction/_instances-0-gw3/node1/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_concurrent_queries_for_all_users_restriction/_instances-0-gw3/.env --project-name roottestconcurrentqueriesforallusersrestriction-gw3 --file /ClickHouse/tests/integration/test_concurrent_queries_for_all_users_restriction/_instances-0-gw3/node1/docker-compose.yml down --volumes] Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Skipping detectLanguage Checking formatRow('foo') Executing query select formatRow('foo') on backward Skipping formatRow Checking countMatchesCaseInsensitive('foo') Executing query select countMatchesCaseInsensitive('foo') on backward Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Skipping countMatchesCaseInsensitive Checking toFloat32OrNull('foo') Executing query select toFloat32OrNull('foo') on backward Executing query SYSTEM FLUSH LOGS on node1 Executing query select toFloat32OrNull('foo') on upstream Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 OK toFloat32OrNull Checking h3IndexesAreNeighbors('foo') Executing query select h3IndexesAreNeighbors('foo') on backward Skipping h3IndexesAreNeighbors Checking tupleDivideByNumber('foo') Executing query select tupleDivideByNumber('foo') on backward Executing query SELECT merge_algorithm, part_type FROM system.part_log WHERE event_type = 'MergeParts' AND table = 't_vertical_merges' AND part_name = 'all_0_1_1'; on node2 Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Skipping tupleDivideByNumber Checking caseWithoutExpression('foo') Executing query select caseWithoutExpression('foo') on backward Executing query SELECT merge_algorithm, part_type FROM system.part_log WHERE event_type = 'MergeParts' AND table = 't_vertical_merges' AND part_name = 'all_0_1_1'; on node1 Skipping caseWithoutExpression Checking fromUnixTimestamp64Milli('foo') Executing query select fromUnixTimestamp64Milli('foo') on backward Executing query ALTER TABLE t_vertical_merges MODIFY SETTING allow_vertical_merges_from_compact_to_wide_parts = 1 on node2 test_backup_restore_on_cluster/test.py::test_table_with_parts_in_queue_considered_non_empty Executing query CREATE DATABASE mydb ON CLUSTER 'cluster' ENGINE=Replicated('/clickhouse/path/','{shard}','{replica}') on node1 Executing query select 20 on node Stderr: Container roottestconcurrentqueriesforallusersrestriction-gw3-node1-1 Stopping Stderr: Container roottestconcurrentqueriesforallusersrestriction-gw3-node1-1 Stopped Stderr: Container roottestconcurrentqueriesforallusersrestriction-gw3-node1-1 Removing Stderr: Container roottestconcurrentqueriesforallusersrestriction-gw3-node1-1 Removed Stderr: Network roottestconcurrentqueriesforallusersrestriction-gw3_default Removing Stderr: Network roottestconcurrentqueriesforallusersrestriction-gw3_default Removed Cleanup called Skipping fromUnixTimestamp64Milli Checking randLogNormal('foo') Executing query select randLogNormal('foo') on backward Docker networks for project roottestconcurrentqueriesforallusersrestriction-gw3 are NETWORK ID NAME DRIVER SCOPE Executing query INSERT INTO t_vertical_merges SELECT number * 3, number, number FROM numbers(60) on node2 Docker containers for project roottestconcurrentqueriesforallusersrestriction-gw3 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestconcurrentqueriesforallusersrestriction-gw3 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestconcurrentqueriesforallusersrestriction-gw3-.*-1$' --format '{{.ID}}:{{.Names}}'] Skipping randLogNormal Checking bar('foo') Executing query select bar('foo') on backward Unstopped containers: {} No running containers for project: roottestconcurrentqueriesforallusersrestriction-gw3 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Executing query OPTIMIZE TABLE t_vertical_merges FINAL on node2 Skipping bar Checking fromModifiedJulianDayOrNull('foo') Executing query select fromModifiedJulianDayOrNull('foo') on backward Stdout:1 Volumes pruned: 1 Executing query CREATE TABLE mydb.test (`x` UInt32) ENGINE = ReplicatedMergeTree ORDER BY x on node1 Skipping fromModifiedJulianDayOrNull Checking bitmaskToList('foo') Executing query select bitmaskToList('foo') on backward Executing query SYSTEM SYNC REPLICA t_vertical_merges on node1 Skipping bitmaskToList Checking shardCount('foo') Executing query select shardCount('foo') on backward Executing query SYSTEM FLUSH LOGS on node2 Skipping shardCount Checking shardNum('foo') Executing query select shardNum('foo') on backward Skipping shardNum Checking readWKTRing('foo') Executing query select readWKTRing('foo') on backward Executing query INSERT INTO mydb.test SELECT number AS x FROM numbers(10000000) on node1 Executing query SYSTEM FLUSH LOGS on node1 Skipping readWKTRing Checking caseWithExpression('foo') Executing query select caseWithExpression('foo') on backward run container_id:roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node1-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "CHECKSUM_DOESNT_MATCH" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true'] Command:[docker exec roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node1-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "CHECKSUM_DOESNT_MATCH" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true] Executing query select 20 on node Skipping caseWithExpression Checking geoDistance('foo') Executing query select geoDistance('foo') on backward run container_id:roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node2-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "CHECKSUM_DOESNT_MATCH" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true'] Command:[docker exec roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node2-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "CHECKSUM_DOESNT_MATCH" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true] Skipping geoDistance Checking greatCircleDistance('foo') Executing query select greatCircleDistance('foo') on backward Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.04.02 03:35:18.169618 [ 588 ] {2ea20ad1-790a-439a-b041-60b324c2164e::all_0_1_1} 2ea20ad1-790a-439a-b041-60b324c2164e::all_0_1_1 (MergeFromLogEntryTask): Code: 40. DB::Exception: Different number of files: 3 compressed (expected 3) and 6 uncompressed ones (expected 5). (CHECKSUM_DOESNT_MATCH) (version 25.2.2.96963.altinityantalya.96963 (official build)). Data after merge is not byte-identical to data on another replicas. There could be several reasons: 1. Using newer version of compression library after server update. 2. Using another compression method. 3. Non-deterministic compression algorithm (highly unlikely). 4. Non-deterministic merge algorithm due to logical error in code. 5. Data corruption in memory due to bug in code. 6. Data corruption in memory due to hardware issue. 7. Manual modification of source data after server startup. 8. Manual modification of checksums stored in ZooKeeper. 9. Part format related settings like 'enable_mixed_granularity_parts' are different on different replicas. We will download merged part from replica to force byte-identical result. List of files in local parts: Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.04.02 03:35:19.061972 [ 600 ] {2ea20ad1-790a-439a-b041-60b324c2164e::all_0_2_2} 2ea20ad1-790a-439a-b041-60b324c2164e::all_0_2_2 (MergeFromLogEntryTask): Code: 40. DB::Exception: Different number of files: 3 compressed (expected 3) and 6 uncompressed ones (expected 5). (CHECKSUM_DOESNT_MATCH) (version 25.2.2.96963.altinityantalya.96963 (official build)). Data after merge is not byte-identical to data on another replicas. There could be several reasons: 1. Using newer version of compression library after server update. 2. Using another compression method. 3. Non-deterministic compression algorithm (highly unlikely). 4. Non-deterministic merge algorithm due to logical error in code. 5. Data corruption in memory due to bug in code. 6. Data corruption in memory due to hardware issue. 7. Manual modification of source data after server startup. 8. Manual modification of checksums stored in ZooKeeper. 9. Part format related settings like 'enable_mixed_granularity_parts' are different on different replicas. We will download merged part from replica to force byte-identical result. List of files in local parts: Executing query SELECT merge_algorithm, part_type FROM system.part_log WHERE event_type = 'MergeParts' AND table = 't_vertical_merges' AND part_name = 'all_0_2_2'; on node2 Skipping greatCircleDistance Checking IPv6StringToNum('foo') Executing query select IPv6StringToNum('foo') on backward Executing query BACKUP DATABASE mydb TO Disk('backups', '30') on node1 Executing query SELECT merge_algorithm, part_type FROM system.part_log WHERE event_type = 'MergeParts' AND table = 't_vertical_merges' AND part_name = 'all_0_2_2'; on node1 Skipping IPv6StringToNum Checking dictGetInt64('foo') Executing query select dictGetInt64('foo') on backward run container_id:roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node1-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill -15 clickhouse'] Command:[docker exec -u root roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node1-1 bash -c pkill -15 clickhouse] run container_id:roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse server' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node1-1 bash -c ps ax | grep 'clickhouse server' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query DROP DATABASE mydb SYNC on node1 Skipping dictGetInt64 Checking greatCircleAngle('foo') Executing query select greatCircleAngle('foo') on backward Stdout:8 Executing query RESTORE DATABASE mydb FROM Disk('backups', '30') on node1 Skipping greatCircleAngle Checking toUnixTimestamp('foo') Executing query select toUnixTimestamp('foo') on backward Skipping toUnixTimestamp Checking h3GetRes0Indexes('foo') Executing query select h3GetRes0Indexes('foo') on backward Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw0] PASSED test_backup_restore_on_cluster/test.py::test_table_with_parts_in_queue_considered_non_empty Skipping h3GetRes0Indexes Checking toIntervalMonth('foo') Executing query select toIntervalMonth('foo') on backward Skipping toIntervalMonth Checking defaultProfiles('foo') Executing query select defaultProfiles('foo') on backward run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Current start attempt failed. Will kill 2020 just in case. run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:True cmd: ['bash', '-c', 'kill -9 2020'] Command:[docker exec -u root roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c kill -9 2020] Skipping defaultProfiles Checking arrayReverse('foo') Executing query select arrayReverse('foo') on backward Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Stderr:bash: line 0: kill: (2020) - No such process Exitcode:1 Skipping arrayReverse Checking now64('foo') Executing query select now64('foo') on backward Skipping now64 Checking reverseDNSQuery('foo') Executing query select reverseDNSQuery('foo') on backward Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Skipping reverseDNSQuery Checking extractAllGroupsHorizontal('foo') Executing query select extractAllGroupsHorizontal('foo') on backward run container_id:roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse server' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node1-1 bash -c ps ax | grep 'clickhouse server' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:8 Skipping extractAllGroupsHorizontal Checking h3ExactEdgeLengthKm('foo') Executing query select h3ExactEdgeLengthKm('foo') on backward Skipping h3ExactEdgeLengthKm Checking arrayFilter('foo') Executing query select arrayFilter('foo') on backward Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Skipping arrayFilter Checking defaultRoles('foo') Executing query select defaultRoles('foo') on backward Skipping defaultRoles Checking toQuarter('foo') Executing query select toQuarter('foo') on backward Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Skipping toQuarter Checking countMatches('foo') Executing query select countMatches('foo') on backward run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] test_backup_restore_on_cluster/test.py::test_tables_dependency Executing query CREATE DATABASE mydb ON CLUSTER 'cluster3' on node1 Skipping countMatches Checking enabledRoles('foo') Executing query select enabledRoles('foo') on backward No clickhouse process running. Start new one. http://localhost:None "POST /v1.46/containers/roottestbackwardcompatibilityconvertordinary-gw5-node-1/exec HTTP/1.1" 201 74 http://localhost:None "POST /v1.46/exec/db5caf1aa669b5e53001fc922131494041a4c5d7cea76fa77acaffe114cd6808/start HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/exec/db5caf1aa669b5e53001fc922131494041a4c5d7cea76fa77acaffe114cd6808/json HTTP/1.1" 200 586 Skipping enabledRoles Checking currentRoles('foo') Executing query select currentRoles('foo') on backward Connection dropped: socket connection error: No route to host Connection dropped: socket connection error: No route to host Skipping currentRoles Checking dateTimeToSnowflake('foo') Executing query select dateTimeToSnowflake('foo') on backward Executing query CREATE TABLE mydb.src ON CLUSTER 'cluster' (x Int64, y String) ENGINE=MergeTree ORDER BY tuple() on node1 run container_id:roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse server' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node1-1 bash -c ps ax | grep 'clickhouse server' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Skipping dateTimeToSnowflake Checking cutToFirstSignificantSubdomain('foo') Executing query select cutToFirstSignificantSubdomain('foo') on backward Stdout:8 Executing query select cutToFirstSignificantSubdomain('foo') on upstream Executing query CREATE DICTIONARY mydb.dict ON CLUSTER 'cluster' (x Int64, y String) PRIMARY KEY x SOURCE(CLICKHOUSE(HOST 'localhost' PORT tcpPort() DB 'mydb' TABLE 'src')) LAYOUT(FLAT()) LIFETIME(0) on node1 OK cutToFirstSignificantSubdomain Checking arrayEnumerateUniqRanked('foo') Executing query select arrayEnumerateUniqRanked('foo') on backward Executing query CREATE TABLE mydb.dist1 (x Int64) ENGINE=Distributed('cluster', 'mydb', 'src') on node1 Skipping arrayEnumerateUniqRanked Checking CRC64('foo') Executing query select CRC64('foo') on backward Executing query CREATE TABLE mydb.dist2 (x Int64) ENGINE=Distributed(cluster, 'mydb', 'src') on node3 Executing query select CRC64('foo') on upstream Executing query CREATE TABLE mydb.clusterfunc1 AS cluster('cluster', 'mydb.src') on node1 OK CRC64 Checking emptyArrayDate('foo') Executing query select emptyArrayDate('foo') on backward Executing query CREATE TABLE mydb.clusterfunc2 AS cluster(cluster, mydb.src) on node1 Skipping emptyArrayDate Checking CRC32IEEE('foo') Executing query select CRC32IEEE('foo') on backward Executing query CREATE TABLE mydb.clusterfunc3 AS cluster(cluster, 'mydb', 'src') on node1 Executing query select CRC32IEEE('foo') on upstream Executing query CREATE TABLE mydb.clusterfunc4 AS cluster(cluster, dictionary(mydb.dict)) on node1 run container_id:roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse server' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node1-1 bash -c ps ax | grep 'clickhouse server' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] OK CRC32IEEE Checking erf('foo') Executing query select erf('foo') on backward Stdout:8 Executing query CREATE TABLE mydb.clusterfunc5 AS clusterAllReplicas(cluster, dictionary(mydb.dict)) on node1 Skipping erf Checking trimBoth('foo') Executing query select trimBoth('foo') on backward Executing query CREATE TABLE mydb.clusterfunc6 AS cluster('cluster', 'mydb.src') on node3 Executing query select trimBoth('foo') on upstream Executing query CREATE TABLE mydb.clusterfunc7 AS cluster(cluster, mydb.src) on node3 OK trimBoth Checking reinterpretAsUInt8('foo') Executing query select reinterpretAsUInt8('foo') on backward Executing query CREATE TABLE mydb.clusterfunc8 AS cluster(cluster, 'mydb', 'src') on node3 run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select reinterpretAsUInt8('foo') on upstream Executing query CREATE TABLE mydb.clusterfunc9 AS cluster(cluster, dictionary(mydb.dict)) on node3 Stdout:2795 Clickhouse process running. run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] OK reinterpretAsUInt8 Checking atanh('foo') Executing query select atanh('foo') on backward Executing query CREATE TABLE mydb.clusterfunc10 AS clusterAllReplicas(cluster, dictionary(mydb.dict)) on node3 Stdout:2795 Executing query select 20 on node Skipping atanh Checking encodeURLComponent('foo') Executing query select encodeURLComponent('foo') on backward Executing query BACKUP DATABASE mydb ON CLUSTER 'cluster3' TO Disk('backups', '31') on node3 Executing query select encodeURLComponent('foo') on upstream run container_id:roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse server' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node1-1 bash -c ps ax | grep 'clickhouse server' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] OK encodeURLComponent Checking wordShingleMinHashArgCaseInsensitiveUTF8('foo') Executing query select wordShingleMinHashArgCaseInsensitiveUTF8('foo') on backward Stdout:8 Executing query select wordShingleMinHashArgCaseInsensitiveUTF8('foo') on upstream Executing query DROP DATABASE mydb on node3 OK wordShingleMinHashArgCaseInsensitiveUTF8 Checking h3GetResolution('foo') Executing query select h3GetResolution('foo') on backward Executing query RESTORE DATABASE mydb ON CLUSTER 'cluster3' FROM Disk('backups', '31') on node3 Skipping h3GetResolution Checking ngramSearch('foo') Executing query select ngramSearch('foo') on backward Skipping ngramSearch Checking lessOrEquals('foo') Executing query select lessOrEquals('foo') on backward Skipping lessOrEquals Checking subtractQuarters('foo') Executing query select subtractQuarters('foo') on backward Executing query SYSTEM FLUSH LOGS ON CLUSTER 'cluster3' on node3 Skipping subtractQuarters Checking fromUnixTimestamp64Nano('foo') Executing query select fromUnixTimestamp64Nano('foo') on backward Executing query select 20 on node Skipping fromUnixTimestamp64Nano Checking toUUIDOrZero('foo') Executing query select toUUIDOrZero('foo') on backward Executing query select toUUIDOrZero('foo') on upstream run container_id:roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse server' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node1-1 bash -c ps ax | grep 'clickhouse server' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] OK toUUIDOrZero Checking h3GetUnidirectionalEdge('foo') Executing query select h3GetUnidirectionalEdge('foo') on backward run container_id:roottestbackuprestoreoncluster-gw0-node1-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.src has no dependencies (level 0)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true'] Command:[docker exec roottestbackuprestoreoncluster-gw0-node1-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.src has no dependencies (level 0)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true] Stdout:8 Skipping h3GetUnidirectionalEdge Checking positiveModulo('foo') Executing query select positiveModulo('foo') on backward Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.04.02 03:35:22.125417 [ 699 ] {f4d70768-5a11-402a-a74a-a5449414da97} RestorerFromBackup: Table mydb.src has no dependencies (level 0) run container_id:roottestbackuprestoreoncluster-gw0-node1-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.dict has 1 dependencies: mydb.src (level 1)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true'] Command:[docker exec roottestbackuprestoreoncluster-gw0-node1-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.dict has 1 dependencies: mydb.src (level 1)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true] Skipping positiveModulo Checking JSONHas('foo') Executing query select JSONHas('foo') on backward Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.04.02 03:35:22.125430 [ 699 ] {f4d70768-5a11-402a-a74a-a5449414da97} RestorerFromBackup: Table mydb.dict has 1 dependencies: mydb.src (level 1) run container_id:roottestbackuprestoreoncluster-gw0-node1-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.dist1 has 1 dependencies: mydb.src (level 1)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true'] Command:[docker exec roottestbackuprestoreoncluster-gw0-node1-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.dist1 has 1 dependencies: mydb.src (level 1)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true] Executing query select JSONHas('foo') on upstream Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.04.02 03:35:22.125448 [ 699 ] {f4d70768-5a11-402a-a74a-a5449414da97} RestorerFromBackup: Table mydb.dist1 has 1 dependencies: mydb.src (level 1) run container_id:roottestbackuprestoreoncluster-gw0-node1-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.clusterfunc1 has 1 dependencies: mydb.src (level 1)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true'] Command:[docker exec roottestbackuprestoreoncluster-gw0-node1-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.clusterfunc1 has 1 dependencies: mydb.src (level 1)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true] OK JSONHas Checking getServerPort('foo') Executing query select getServerPort('foo') on backward Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.04.02 03:35:22.125440 [ 699 ] {f4d70768-5a11-402a-a74a-a5449414da97} RestorerFromBackup: Table mydb.clusterfunc1 has 1 dependencies: mydb.src (level 1) run container_id:roottestbackuprestoreoncluster-gw0-node1-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.clusterfunc2 has 1 dependencies: mydb.src (level 1)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true'] Command:[docker exec roottestbackuprestoreoncluster-gw0-node1-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.clusterfunc2 has 1 dependencies: mydb.src (level 1)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true] Skipping getServerPort Checking runningDifference('foo') Executing query select runningDifference('foo') on backward Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.04.02 03:35:22.125458 [ 699 ] {f4d70768-5a11-402a-a74a-a5449414da97} RestorerFromBackup: Table mydb.clusterfunc2 has 1 dependencies: mydb.src (level 1) run container_id:roottestbackuprestoreoncluster-gw0-node1-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.clusterfunc3 has 1 dependencies: mydb.src (level 1)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true'] Command:[docker exec roottestbackuprestoreoncluster-gw0-node1-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.clusterfunc3 has 1 dependencies: mydb.src (level 1)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true] Skipping runningDifference Checking arrayPushBack('foo') Executing query select arrayPushBack('foo') on backward Executing query select 20 on node Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.04.02 03:35:22.125453 [ 699 ] {f4d70768-5a11-402a-a74a-a5449414da97} RestorerFromBackup: Table mydb.clusterfunc3 has 1 dependencies: mydb.src (level 1) run container_id:roottestbackuprestoreoncluster-gw0-node1-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.clusterfunc4 has 1 dependencies: mydb.dict (level 2)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true'] Command:[docker exec roottestbackuprestoreoncluster-gw0-node1-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.clusterfunc4 has 1 dependencies: mydb.dict (level 2)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true] Skipping arrayPushBack Checking hopEnd('foo') Executing query select hopEnd('foo') on backward Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.04.02 03:35:22.125476 [ 699 ] {f4d70768-5a11-402a-a74a-a5449414da97} RestorerFromBackup: Table mydb.clusterfunc4 has 1 dependencies: mydb.dict (level 2) run container_id:roottestbackuprestoreoncluster-gw0-node1-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.clusterfunc5 has 1 dependencies: mydb.dict (level 2)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true'] Command:[docker exec roottestbackuprestoreoncluster-gw0-node1-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.clusterfunc5 has 1 dependencies: mydb.dict (level 2)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true] Skipping hopEnd Checking arrayAvg('foo') Executing query select arrayAvg('foo') on backward Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.04.02 03:35:22.125464 [ 699 ] {f4d70768-5a11-402a-a74a-a5449414da97} RestorerFromBackup: Table mydb.clusterfunc5 has 1 dependencies: mydb.dict (level 2) run container_id:roottestbackuprestoreoncluster-gw0-node2-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.src has no dependencies (level 0)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true'] Command:[docker exec roottestbackuprestoreoncluster-gw0-node2-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.src has no dependencies (level 0)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true] run container_id:roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse server' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node1-1 bash -c ps ax | grep 'clickhouse server' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Skipping arrayAvg Checking degrees('foo') Executing query select degrees('foo') on backward Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.04.02 03:35:22.122398 [ 1559 ] {3fbb6ec3-242d-4ab3-9d46-2585fc073969} RestorerFromBackup: Table mydb.src has no dependencies (level 0) run container_id:roottestbackuprestoreoncluster-gw0-node2-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.dict has 1 dependencies: mydb.src (level 1)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true'] Command:[docker exec roottestbackuprestoreoncluster-gw0-node2-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.dict has 1 dependencies: mydb.src (level 1)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true] Stdout:8 Skipping degrees Checking concatWithSeparator('foo') Executing query select concatWithSeparator('foo') on backward Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.04.02 03:35:22.122412 [ 1559 ] {3fbb6ec3-242d-4ab3-9d46-2585fc073969} RestorerFromBackup: Table mydb.dict has 1 dependencies: mydb.src (level 1) run container_id:roottestbackuprestoreoncluster-gw0-node3-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.dist2 has no dependencies (level 0)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true'] Command:[docker exec roottestbackuprestoreoncluster-gw0-node3-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.dist2 has no dependencies (level 0)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true] Executing query select concatWithSeparator('foo') on upstream Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.04.02 03:35:22.126227 [ 685 ] {9597f39a-a280-4803-b1c4-98cec5f96795} RestorerFromBackup: Table mydb.dist2 has no dependencies (level 0) run container_id:roottestbackuprestoreoncluster-gw0-node3-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.clusterfunc6 has no dependencies (level 0)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true'] Command:[docker exec roottestbackuprestoreoncluster-gw0-node3-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.clusterfunc6 has no dependencies (level 0)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true] OK concatWithSeparator Checking dictGetDateTime('foo') Executing query select dictGetDateTime('foo') on backward Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.04.02 03:35:22.126235 [ 685 ] {9597f39a-a280-4803-b1c4-98cec5f96795} RestorerFromBackup: Table mydb.clusterfunc6 has no dependencies (level 0) run container_id:roottestbackuprestoreoncluster-gw0-node3-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.clusterfunc7 has no dependencies (level 0)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true'] Command:[docker exec roottestbackuprestoreoncluster-gw0-node3-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.clusterfunc7 has no dependencies (level 0)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true] Skipping dictGetDateTime Checking h3ToGeoBoundary('foo') Executing query select h3ToGeoBoundary('foo') on backward Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.04.02 03:35:22.126231 [ 685 ] {9597f39a-a280-4803-b1c4-98cec5f96795} RestorerFromBackup: Table mydb.clusterfunc7 has no dependencies (level 0) run container_id:roottestbackuprestoreoncluster-gw0-node3-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.clusterfunc8 has no dependencies (level 0)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true'] Command:[docker exec roottestbackuprestoreoncluster-gw0-node3-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.clusterfunc8 has no dependencies (level 0)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true] Skipping h3ToGeoBoundary Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.04.02 03:35:22.126243 [ 685 ] {9597f39a-a280-4803-b1c4-98cec5f96795} RestorerFromBackup: Table mydb.clusterfunc8 has no dependencies (level 0) Checking multiSearchAnyUTF8('foo') run container_id:roottestbackuprestoreoncluster-gw0-node3-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.clusterfunc9 has no dependencies (level 0)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true'] Executing query select multiSearchAnyUTF8('foo') on backward Command:[docker exec roottestbackuprestoreoncluster-gw0-node3-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.clusterfunc9 has no dependencies (level 0)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true] Executing query select 20 on node Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.04.02 03:35:22.126220 [ 685 ] {9597f39a-a280-4803-b1c4-98cec5f96795} RestorerFromBackup: Table mydb.clusterfunc9 has no dependencies (level 0) run container_id:roottestbackuprestoreoncluster-gw0-node3-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.clusterfunc10 has no dependencies (level 0)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true'] Command:[docker exec roottestbackuprestoreoncluster-gw0-node3-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.clusterfunc10 has no dependencies (level 0)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true] Skipping multiSearchAnyUTF8 Checking mapApply('foo') Executing query select mapApply('foo') on backward Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.04.02 03:35:22.126239 [ 685 ] {9597f39a-a280-4803-b1c4-98cec5f96795} RestorerFromBackup: Table mydb.clusterfunc10 has no dependencies (level 0) Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw0] PASSED test_backup_restore_on_cluster/test.py::test_tables_dependency Skipping mapApply Checking atan('foo') Executing query select atan('foo') on backward run container_id:roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse server' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node1-1 bash -c ps ax | grep 'clickhouse server' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Skipping atan Checking formatDateTimeInJodaSyntax('foo') Executing query select formatDateTimeInJodaSyntax('foo') on backward Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Stdout:8 Skipping formatDateTimeInJodaSyntax Checking IPv4StringToNum('foo') Executing query select IPv4StringToNum('foo') on backward Skipping IPv4StringToNum Checking positionUTF8('foo') Executing query select positionUTF8('foo') on backward Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Skipping positionUTF8 Checking hasToken('foo') Executing query select hasToken('foo') on backward Skipping hasToken Checking hypot('foo') Executing query select hypot('foo') on backward Connection dropped: socket connection error: No route to host Connection dropped: socket connection error: No route to host Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Skipping hypot Checking protocol('foo') Executing query select protocol('foo') on backward Executing query select 20 on node Executing query select protocol('foo') on upstream Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 OK protocol Checking if('foo') Executing query select if('foo') on backward Command:[docker compose --env-file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/.env --project-name roottestbackuprestoreoncluster-gw0 --file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/node2/docker-compose.yml --file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/node3/docker-compose.yml stop --timeout 20] Skipping if Checking toIntervalSecond('foo') Executing query select toIntervalSecond('foo') on backward run container_id:roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse server' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node1-1 bash -c ps ax | grep 'clickhouse server' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Skipping toIntervalSecond Checking kostikConsistentHash('foo') Executing query select kostikConsistentHash('foo') on backward Stdout:8 Skipping kostikConsistentHash Checking regionToContinent('foo') Executing query select regionToContinent('foo') on backward Skipping regionToContinent Checking makeDate32('foo') Executing query select makeDate32('foo') on backward Skipping makeDate32 Checking ilike('foo') Executing query select ilike('foo') on backward Stderr: Container roottestbrokenpartduringmerge-gw8-node1-1 Stopping Stderr: Container roottestbrokenpartduringmerge-gw8-node1-1 Stopped Stderr: Container roottestbrokenpartduringmerge-gw8-zoo2-1 Stopping Stderr: Container roottestbrokenpartduringmerge-gw8-zoo3-1 Stopping Stderr: Container roottestbrokenpartduringmerge-gw8-zoo1-1 Stopping Stderr: Container roottestbrokenpartduringmerge-gw8-zoo2-1 Stopped Stderr: Container roottestbrokenpartduringmerge-gw8-zoo3-1 Stopped Stderr: Container roottestbrokenpartduringmerge-gw8-zoo1-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_broken_part_during_merge/_instances-0-gw8/node1/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_broken_part_during_merge/_instances-0-gw8/node1/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_broken_part_during_merge/_instances-0-gw8/.env --project-name roottestbrokenpartduringmerge-gw8 --file /ClickHouse/tests/integration/test_broken_part_during_merge/_instances-0-gw8/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml down --volumes] Executing query select 20 on node Skipping ilike Checking toUInt64OrZero('foo') Executing query select toUInt64OrZero('foo') on backward Executing query select toUInt64OrZero('foo') on upstream run container_id:roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse server' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node1-1 bash -c ps ax | grep 'clickhouse server' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] OK toUInt64OrZero Checking dictGetUInt16('foo') Executing query select dictGetUInt16('foo') on backward Stdout:8 Skipping dictGetUInt16 Checking globalNotNullInIgnoreSet('foo') Executing query select globalNotNullInIgnoreSet('foo') on backward Skipping globalNotNullInIgnoreSet Checking h3GetUnidirectionalEdgeBoundary('foo') Executing query select h3GetUnidirectionalEdgeBoundary('foo') on backward Skipping h3GetUnidirectionalEdgeBoundary Checking globalNullInIgnoreSet('foo') Executing query select globalNullInIgnoreSet('foo') on backward Skipping globalNullInIgnoreSet Checking hasAll('foo') Executing query select hasAll('foo') on backward Skipping hasAll Checking toTime('foo') Executing query select toTime('foo') on backward Stderr: Container roottestbrokenpartduringmerge-gw8-node1-1 Stopping Stderr: Container roottestbrokenpartduringmerge-gw8-node1-1 Stopped Stderr: Container roottestbrokenpartduringmerge-gw8-node1-1 Removing Stderr: Container roottestbrokenpartduringmerge-gw8-node1-1 Removed Stderr: Container roottestbrokenpartduringmerge-gw8-zoo2-1 Stopping Stderr: Container roottestbrokenpartduringmerge-gw8-zoo3-1 Stopping Stderr: Container roottestbrokenpartduringmerge-gw8-zoo1-1 Stopping Stderr: Container roottestbrokenpartduringmerge-gw8-zoo3-1 Stopped Stderr: Container roottestbrokenpartduringmerge-gw8-zoo3-1 Removing Stderr: Container roottestbrokenpartduringmerge-gw8-zoo2-1 Stopped Stderr: Container roottestbrokenpartduringmerge-gw8-zoo2-1 Removing Stderr: Container roottestbrokenpartduringmerge-gw8-zoo1-1 Stopped Stderr: Container roottestbrokenpartduringmerge-gw8-zoo1-1 Removing Stderr: Container roottestbrokenpartduringmerge-gw8-zoo1-1 Removed Stderr: Container roottestbrokenpartduringmerge-gw8-zoo2-1 Removed Stderr: Container roottestbrokenpartduringmerge-gw8-zoo3-1 Removed Stderr: Network roottestbrokenpartduringmerge-gw8_default Removing Stderr: Network roottestbrokenpartduringmerge-gw8_default Removed Cleanup called Executing query select 20 on node Docker networks for project roottestbrokenpartduringmerge-gw8 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestbrokenpartduringmerge-gw8 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Skipping toTime Checking globalNotInIgnoreSet('foo') Executing query select globalNotInIgnoreSet('foo') on backward Docker volumes for project roottestbrokenpartduringmerge-gw8 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestbrokenpartduringmerge-gw8-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestbrokenpartduringmerge-gw8 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Skipping globalNotInIgnoreSet Checking splitByChar('foo') Executing query select splitByChar('foo') on backward Stdout:1 Volumes pruned: 1 test_cgroup_limit/test.py::test_cgroup_cpu_limit Command: docker run --rm --cpus 1 --volume /clickhouse:/usr/bin/clickhouse ubuntu:22.04 sh -c clickhouse local -q "select value from system.settings where name='max_threads'" Unable to find image 'ubuntu:22.04' locally Skipping splitByChar Checking toRelativeDayNum('foo') Executing query select toRelativeDayNum('foo') on backward run container_id:roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse server' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node1-1 bash -c ps ax | grep 'clickhouse server' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Skipping toRelativeDayNum Checking notInIgnoreSet('foo') Executing query select notInIgnoreSet('foo') on backward run container_id:roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse server' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node1-1 bash -c ps ax | grep 'clickhouse server' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Skipping notInIgnoreSet Checking globalInIgnoreSet('foo') Executing query select globalInIgnoreSet('foo') on backward run container_id:roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node1-1 detach:False nothrow:False cmd: ['bash', '-c', 'cp /usr/bin/clickhouse /usr/share/clickhouse_original'] Command:[docker exec -u root roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node1-1 bash -c cp /usr/bin/clickhouse /usr/share/clickhouse_original] Skipping globalInIgnoreSet Checking dictGetUUID('foo') Executing query select dictGetUUID('foo') on backward Skipping dictGetUUID Checking inIgnoreSet('foo') Executing query select inIgnoreSet('foo') on backward Skipping inIgnoreSet Checking arrayConcat('foo') Executing query select arrayConcat('foo') on backward Executing query select 20 on node Skipping arrayConcat Checking globalNotNullIn('foo') Executing query select globalNotNullIn('foo') on backward run container_id:roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node1-1 detach:False nothrow:False cmd: ['bash', '-c', 'cp /usr/share/clickhouse_fresh /usr/bin/clickhouse && chmod 777 /usr/bin/clickhouse'] Command:[docker exec -u root roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node1-1 bash -c cp /usr/share/clickhouse_fresh /usr/bin/clickhouse && chmod 777 /usr/bin/clickhouse] Skipping globalNotNullIn Checking notNullIn('foo') Executing query select notNullIn('foo') on backward Skipping notNullIn Checking globalNullIn('foo') Executing query select globalNullIn('foo') on backward Executing query DROP TABLE IF EXISTS tbl SYNC on node [gw2] PASSED test_backup_restore_new/test_shutdown_wait_backup.py::test_shutdown_wait_backup Skipping globalNullIn Checking dateTime64ToSnowflake('foo') Executing query select dateTime64ToSnowflake('foo') on backward Command:[docker compose --env-file /ClickHouse/tests/integration/test_backup_restore_new/_instances-shutdown_wait_backup-0-gw2/.env --project-name roottestbackuprestorenewshutdownwaitbackup-gw2 --file /ClickHouse/tests/integration/test_backup_restore_new/_instances-shutdown_wait_backup-0-gw2/node/docker-compose.yml stop --timeout 20] Skipping dateTime64ToSnowflake Checking dictGetInt8OrDefault('foo') Executing query select dictGetInt8OrDefault('foo') on backward Skipping dictGetInt8OrDefault Checking addressToLine('foo') Executing query select addressToLine('foo') on backward Skipping addressToLine Checking globalNotIn('foo') Executing query select globalNotIn('foo') on backward Skipping globalNotIn Checking ifNull('foo') Executing query select ifNull('foo') on backward Skipping ifNull Checking normalizeQueryKeepNames('foo') Executing query select normalizeQueryKeepNames('foo') on backward Executing query select 20 on node run container_id:roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "echo 'restart_with_latest_version: From version' && /usr/share/clickhouse_original server --version && echo 'To version' /usr/share/clickhouse_fresh server --version"] Command:[docker exec roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node1-1 bash -c echo 'restart_with_latest_version: From version' && /usr/share/clickhouse_original server --version && echo 'To version' /usr/share/clickhouse_fresh server --version] Stdout:restart_with_latest_version: From version Stdout:ClickHouse server version 23.3.22.3 (official build). Stdout:To version /usr/share/clickhouse_fresh server --version run container_id:roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node1-1 detach:False nothrow:False cmd: ['bash', '-c', 'clickhouse server --config-file=/etc/clickhouse-server/config.xml --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log --daemon -- '] Command:[docker exec -u 0 roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node1-1 bash -c clickhouse server --config-file=/etc/clickhouse-server/config.xml --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log --daemon -- ] Executing query select normalizeQueryKeepNames('foo') on upstream run container_id:roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:486 Executing query select 20 on node1 OK normalizeQueryKeepNames Checking xxHash64('foo') Executing query select xxHash64('foo') on backward Executing query select xxHash64('foo') on upstream Stderr: Container roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 Stopping Stderr: Container roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_backup_restore_new/_instances-shutdown_wait_backup-0-gw2/node/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_backup_restore_new/_instances-shutdown_wait_backup-0-gw2/node/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_backup_restore_new/_instances-shutdown_wait_backup-0-gw2/.env --project-name roottestbackuprestorenewshutdownwaitbackup-gw2 --file /ClickHouse/tests/integration/test_backup_restore_new/_instances-shutdown_wait_backup-0-gw2/node/docker-compose.yml down --volumes] OK xxHash64 Checking in('foo') Executing query select in('foo') on backward Skipping in Checking firstSignificantSubdomain('foo') Executing query select firstSignificantSubdomain('foo') on backward Executing query select firstSignificantSubdomain('foo') on upstream Executing query select 20 on node OK firstSignificantSubdomain Checking h3HexAreaKm2('foo') Executing query select h3HexAreaKm2('foo') on backward Skipping h3HexAreaKm2 Checking connectionId('foo') Executing query select connectionId('foo') on backward Executing query select 20 on node1 Skipping connectionId Checking ngramDistanceCaseInsensitive('foo') Executing query select ngramDistanceCaseInsensitive('foo') on backward run container_id:roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node2-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] Command:[docker exec -u root roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node2-1 bash -c ps -C clickhouse] Skipping ngramDistanceCaseInsensitive Checking startsWith('foo') Executing query select startsWith('foo') on backward Stdout: PID TTY TIME CMD Stdout: 10 ? 00:00:01 clickhouse run container_id:roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node2-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] Command:[docker exec -u root roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node2-1 bash -c pkill clickhouse] run container_id:roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Skipping startsWith Checking initialQueryID('foo') Executing query select initialQueryID('foo') on backward Stderr: Container roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 Stopping Stderr: Container roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 Stopped Stderr: Container roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 Removing Stderr: Container roottestbackuprestorenewshutdownwaitbackup-gw2-node-1 Removed Stderr: Network roottestbackuprestorenewshutdownwaitbackup-gw2_default Removing Stderr: Network roottestbackuprestorenewshutdownwaitbackup-gw2_default Removed Cleanup called Docker networks for project roottestbackuprestorenewshutdownwaitbackup-gw2 are NETWORK ID NAME DRIVER SCOPE Stdout:10 Docker containers for project roottestbackuprestorenewshutdownwaitbackup-gw2 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Skipping initialQueryID Checking murmurHash2_64('foo') Executing query select murmurHash2_64('foo') on backward Docker volumes for project roottestbackuprestorenewshutdownwaitbackup-gw2 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestbackuprestorenewshutdownwaitbackup-gw2-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestbackuprestorenewshutdownwaitbackup-gw2 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:1 Volumes pruned: 1 test_backward_compatibility/test_normalized_count_comparison.py::test_select_aggregate_alias_column Running tests in /ClickHouse/tests/integration/test_backward_compatibility/test_normalized_count_comparison.py Cluster start called. is_up=False Docker networks for project roottestbackwardcompatibilitynormalizedcountcomparison-gw2 are NETWORK ID NAME DRIVER SCOPE Executing query select murmurHash2_64('foo') on upstream Docker containers for project roottestbackwardcompatibilitynormalizedcountcomparison-gw2 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestbackwardcompatibilitynormalizedcountcomparison-gw2 are DRIVER VOLUME NAME Cleanup called Docker networks for project roottestbackwardcompatibilitynormalizedcountcomparison-gw2 are NETWORK ID NAME DRIVER SCOPE run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Docker containers for project roottestbackwardcompatibilitynormalizedcountcomparison-gw2 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES OK murmurHash2_64 Checking h3GetBaseCell('foo') Executing query select h3GetBaseCell('foo') on backward Docker volumes for project roottestbackwardcompatibilitynormalizedcountcomparison-gw2 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestbackwardcompatibilitynormalizedcountcomparison-gw2-.*-1$' --format '{{.ID}}:{{.Names}}'] Current start attempt failed. Will kill 2795 just in case. run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:True cmd: ['bash', '-c', 'kill -9 2795'] Command:[docker exec -u root roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c kill -9 2795] Unstopped containers: {} No running containers for project: roottestbackwardcompatibilitynormalizedcountcomparison-gw2 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Stderr:bash: line 0: kill: (2795) - No such process Exitcode:1 Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Skipping h3GetBaseCell Checking bitmapHasAll('foo') Executing query select bitmapHasAll('foo') on backward Stdout:1 Volumes pruned: 1 Setup directory for instance: node1 Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files [] to /ClickHouse/tests/integration/test_backward_compatibility/_instances-normalized_count_comparison-0-gw2/node1/configs/config.d Setup database dir /ClickHouse/tests/integration/test_backward_compatibility/_instances-normalized_count_comparison-0-gw2/node1/database Setup logs dir /ClickHouse/tests/integration/test_backward_compatibility/_instances-normalized_count_comparison-0-gw2/node1/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Setup directory for instance: node2 Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files [] to /ClickHouse/tests/integration/test_backward_compatibility/_instances-normalized_count_comparison-0-gw2/node2/configs/config.d Setup database dir /ClickHouse/tests/integration/test_backward_compatibility/_instances-normalized_count_comparison-0-gw2/node2/database Setup logs dir /ClickHouse/tests/integration/test_backward_compatibility/_instances-normalized_count_comparison-0-gw2/node2/logs Entrypoint cmd: bash -c "trap 'pkill tail' INT TERM; clickhouse server --config-file=/etc/clickhouse-server/config.xml --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log --daemon -- ; coproc tail -f /dev/null; wait $$!" Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw'} stored in /ClickHouse/tests/integration/test_backward_compatibility/_instances-normalized_count_comparison-0-gw2/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_backward_compatibility/_instances-normalized_count_comparison-0-gw2/.env --project-name roottestbackwardcompatibilitynormalizedcountcomparison-gw2 --file /ClickHouse/tests/integration/test_backward_compatibility/_instances-normalized_count_comparison-0-gw2/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_backward_compatibility/_instances-normalized_count_comparison-0-gw2/node2/docker-compose.yml pull] Skipping bitmapHasAll Checking base58Decode('foo') Executing query select base58Decode('foo') on backward Executing query select base58Decode('foo') on upstream OK base58Decode Checking regionToTopContinent('foo') Executing query select regionToTopContinent('foo') on backward Skipping regionToTopContinent Checking cityHash64('foo') Executing query select cityHash64('foo') on backward Executing query select cityHash64('foo') on upstream OK cityHash64 Checking intDiv('foo') Executing query select intDiv('foo') on backward Skipping intDiv Checking intExp10('foo') Executing query select intExp10('foo') on backward Skipping intExp10 Checking errorCodeToName('foo') Executing query select errorCodeToName('foo') on backward Skipping errorCodeToName Checking fromUnixTimestampInJodaSyntax('foo') Executing query select fromUnixTimestampInJodaSyntax('foo') on backward run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Skipping fromUnixTimestampInJodaSyntax Checking log('foo') Executing query select log('foo') on backward No clickhouse process running. Start new one. http://localhost:None "POST /v1.46/containers/roottestbackwardcompatibilityconvertordinary-gw5-node-1/exec HTTP/1.1" 201 74 http://localhost:None "POST /v1.46/exec/046bb33849dc8a163300a77b068f7cf00a3fd90757331f50eafd7330ff4ca02f/start HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/exec/046bb33849dc8a163300a77b068f7cf00a3fd90757331f50eafd7330ff4ca02f/json HTTP/1.1" 200 586 Skipping log Checking randomFixedString('foo') Executing query select randomFixedString('foo') on backward Skipping randomFixedString Checking toWeek('foo') Executing query select toWeek('foo') on backward Skipping toWeek Checking toStartOfHour('foo') Executing query select toStartOfHour('foo') on backward run container_id:roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Skipping toStartOfHour Checking isDecimalOverflow('foo') Executing query select isDecimalOverflow('foo') on backward Stdout:10 Skipping isDecimalOverflow Checking encodeURLFormComponent('foo') Executing query select encodeURLFormComponent('foo') on backward Executing query select encodeURLFormComponent('foo') on upstream OK encodeURLFormComponent Checking getSizeOfEnumType('foo') Executing query select getSizeOfEnumType('foo') on backward Skipping getSizeOfEnumType Checking detectLanguageMixed('foo') Executing query select detectLanguageMixed('foo') on backward Skipping detectLanguageMixed Checking tumble('foo') Executing query select tumble('foo') on backward Skipping tumble Checking toStartOfInterval('foo') Executing query select toStartOfInterval('foo') on backward Skipping toStartOfInterval Checking hasTokenCaseInsensitiveOrNull('foo') Executing query select hasTokenCaseInsensitiveOrNull('foo') on backward Skipping hasTokenCaseInsensitiveOrNull Checking multiFuzzyMatchAnyIndex('foo') Executing query select multiFuzzyMatchAnyIndex('foo') on backward Skipping multiFuzzyMatchAnyIndex Checking isZeroOrNull('foo') Executing query select isZeroOrNull('foo') on backward Skipping isZeroOrNull Checking lower('foo') Executing query select lower('foo') on backward Executing query select lower('foo') on upstream OK lower Checking countEqual('foo') Executing query select countEqual('foo') on backward run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Skipping countEqual Checking JSONArrayLength('foo') Executing query select JSONArrayLength('foo') on backward Stdout:3553 Clickhouse process running. run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:3553 Executing query select 20 on node Executing query select JSONArrayLength('foo') on upstream OK JSONArrayLength Checking asin('foo') Executing query select asin('foo') on backward run container_id:roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Skipping asin Checking left('foo') Executing query select left('foo') on backward Stdout:10 Skipping left Checking runningDifferenceStartingWithFirstValue('foo') Executing query select runningDifferenceStartingWithFirstValue('foo') on backward Skipping runningDifferenceStartingWithFirstValue Checking h3CellAreaM2('foo') Executing query select h3CellAreaM2('foo') on backward Skipping h3CellAreaM2 Checking h3PointDistM('foo') Executing query select h3PointDistM('foo') on backward Skipping h3PointDistM Checking tumbleEnd('foo') Executing query select tumbleEnd('foo') on backward Skipping tumbleEnd Checking isInfinite('foo') Executing query select isInfinite('foo') on backward Executing query select 20 on node Skipping isInfinite Checking lemmatize('foo') Executing query select lemmatize('foo') on backward Skipping lemmatize Checking reverseUTF8('foo') Executing query select reverseUTF8('foo') on backward Executing query select reverseUTF8('foo') on upstream OK reverseUTF8 Checking murmurHash3_64('foo') Executing query select murmurHash3_64('foo') on backward Executing query select murmurHash3_64('foo') on upstream OK murmurHash3_64 Checking rightPad('foo') Executing query select rightPad('foo') on backward Skipping rightPad Checking bitNot('foo') Executing query select bitNot('foo') on backward Executing query select 20 on node Skipping bitNot Checking lengthUTF8('foo') Executing query select lengthUTF8('foo') on backward run container_id:roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:10 Executing query select lengthUTF8('foo') on upstream OK lengthUTF8 Checking toStartOfFiveMinutes('foo') Executing query select toStartOfFiveMinutes('foo') on backward Skipping toStartOfFiveMinutes Checking reinterpretAsUInt128('foo') Executing query select reinterpretAsUInt128('foo') on backward Executing query select reinterpretAsUInt128('foo') on upstream OK reinterpretAsUInt128 Checking lgamma('foo') Executing query select lgamma('foo') on backward Skipping lgamma Checking toInt16('foo') Executing query select toInt16('foo') on backward Executing query select 20 on node Skipping toInt16 Checking reverse('foo') Executing query select reverse('foo') on backward Executing query select reverse('foo') on upstream OK reverse Checking reinterpretAsDate('foo') Executing query select reinterpretAsDate('foo') on backward Executing query select reinterpretAsDate('foo') on upstream OK reinterpretAsDate Checking like('foo') Executing query select like('foo') on backward Skipping like Checking SHA224('foo') Executing query select SHA224('foo') on backward Executing query select SHA224('foo') on upstream OK SHA224 Checking h3kRing('foo') Executing query select h3kRing('foo') on backward Skipping h3kRing Checking log2('foo') Executing query select log2('foo') on backward Executing query select 20 on node run container_id:roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Skipping log2 Checking exp10('foo') Executing query select exp10('foo') on backward run container_id:roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Skipping exp10 Checking toIPv6OrDefault('foo') Executing query select toIPv6OrDefault('foo') on backward No clickhouse process running. Start new one. http://localhost:None "POST /v1.46/containers/roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node2-1/exec HTTP/1.1" 201 74 http://localhost:None "POST /v1.46/exec/39538ac5e6b70742f85440f6bce1fd7062a925fc1f326c6fd46537c1f505761b/start HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/exec/39538ac5e6b70742f85440f6bce1fd7062a925fc1f326c6fd46537c1f505761b/json HTTP/1.1" 200 586 Executing query select toIPv6OrDefault('foo') on upstream OK toIPv6OrDefault Checking toIPv4OrDefault('foo') Executing query select toIPv4OrDefault('foo') on backward Executing query select toIPv4OrDefault('foo') on upstream OK toIPv4OrDefault Checking L2SquaredDistance('foo') Executing query select L2SquaredDistance('foo') on backward Skipping L2SquaredDistance Checking emptyArrayInt64('foo') Executing query select emptyArrayInt64('foo') on backward Skipping emptyArrayInt64 Checking toDecimal256OrDefault('foo') Executing query select toDecimal256OrDefault('foo') on backward Executing query select 20 on node Skipping toDecimal256OrDefault Checking partitionId('foo') Executing query select partitionId('foo') on backward Executing query select partitionId('foo') on upstream OK partitionId Checking roundToExp2('foo') Executing query select roundToExp2('foo') on backward Skipping roundToExp2 Checking mapSubtract('foo') Executing query select mapSubtract('foo') on backward Skipping mapSubtract Checking toDecimal32OrDefault('foo') Executing query select toDecimal32OrDefault('foo') on backward Skipping toDecimal32OrDefault Checking cutToFirstSignificantSubdomainWithWWWRFC('foo') Executing query select cutToFirstSignificantSubdomainWithWWWRFC('foo') on backward Executing query select cutToFirstSignificantSubdomainWithWWWRFC('foo') on upstream OK cutToFirstSignificantSubdomainWithWWWRFC Checking toDateTimeOrDefault('foo') Executing query select toDateTimeOrDefault('foo') on backward Executing query select toDateTimeOrDefault('foo') on upstream Executing query select 20 on node OK toDateTimeOrDefault Checking addMinutes('foo') Executing query select addMinutes('foo') on backward Skipping addMinutes Checking translate('foo') Executing query select translate('foo') on backward run container_id:roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:798 Clickhouse process running. run container_id:roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Skipping translate Checking trimRight('foo') Executing query select trimRight('foo') on backward Stdout:798 Executing query select 20 on node2 Executing query select trimRight('foo') on upstream Executing query ALTER TABLE t_vertical_merges MODIFY SETTING allow_vertical_merges_from_compact_to_wide_parts = 1 on node1 OK trimRight Checking runningConcurrency('foo') Executing query select runningConcurrency('foo') on backward Executing query INSERT INTO t_vertical_merges SELECT number * 4, number, number FROM numbers(60) on node2 Skipping runningConcurrency Checking toDateOrDefault('foo') Executing query select toDateOrDefault('foo') on backward Executing query OPTIMIZE TABLE t_vertical_merges FINAL on node2 Executing query select toDateOrDefault('foo') on upstream Executing query SYSTEM SYNC REPLICA t_vertical_merges on node1 OK toDateOrDefault Checking plus('foo') Executing query select plus('foo') on backward Executing query select 20 on node Executing query SYSTEM FLUSH LOGS on node2 Skipping plus Checking toInt256OrDefault('foo') Executing query select toInt256OrDefault('foo') on backward Executing query select toInt256OrDefault('foo') on upstream Executing query SYSTEM FLUSH LOGS on node1 OK toInt256OrDefault Checking dictGetIPv4('foo') Executing query select dictGetIPv4('foo') on backward Skipping dictGetIPv4 Checking nullIn('foo') Executing query select nullIn('foo') on backward Skipping nullIn Checking toDecimal128OrZero('foo') Executing query select toDecimal128OrZero('foo') on backward run container_id:roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node1-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "CHECKSUM_DOESNT_MATCH" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true'] Command:[docker exec roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node1-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "CHECKSUM_DOESNT_MATCH" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true] Skipping toDecimal128OrZero Checking extractURLParameters('foo') Executing query select extractURLParameters('foo') on backward run container_id:roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node2-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "CHECKSUM_DOESNT_MATCH" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true'] Command:[docker exec roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node2-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "CHECKSUM_DOESNT_MATCH" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true] Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.04.02 03:35:32.547161 [ 1395 ] {2ea20ad1-790a-439a-b041-60b324c2164e::all_0_3_3} 2ea20ad1-790a-439a-b041-60b324c2164e::all_0_3_3 (MergeFromLogEntryTask): Code: 40. DB::Exception: Different number of files: 3 compressed (expected 7) and 6 uncompressed ones (expected 2). (CHECKSUM_DOESNT_MATCH) (version 25.2.2.96963.altinityantalya.96963 (official build)). Data after merge is not byte-identical to data on another replicas. There could be several reasons: 1. Using newer version of compression library after server update. 2. Using another compression method. 3. Non-deterministic compression algorithm (highly unlikely). 4. Non-deterministic merge algorithm due to logical error in code. 5. Data corruption in memory due to bug in code. 6. Data corruption in memory due to hardware issue. 7. Manual modification of source data after server startup. 8. Manual modification of checksums stored in ZooKeeper. 9. Part format related settings like 'enable_mixed_granularity_parts' are different on different replicas. We will download merged part from replica to force byte-identical result. List of files in local parts: run container_id:roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node2-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "Different number of files" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true'] Command:[docker exec roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node2-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "Different number of files" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true] Executing query select extractURLParameters('foo') on upstream Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.04.02 03:35:32.547161 [ 1395 ] {2ea20ad1-790a-439a-b041-60b324c2164e::all_0_3_3} 2ea20ad1-790a-439a-b041-60b324c2164e::all_0_3_3 (MergeFromLogEntryTask): Code: 40. DB::Exception: Different number of files: 3 compressed (expected 7) and 6 uncompressed ones (expected 2). (CHECKSUM_DOESNT_MATCH) (version 25.2.2.96963.altinityantalya.96963 (official build)). Data after merge is not byte-identical to data on another replicas. There could be several reasons: 1. Using newer version of compression library after server update. 2. Using another compression method. 3. Non-deterministic compression algorithm (highly unlikely). 4. Non-deterministic merge algorithm due to logical error in code. 5. Data corruption in memory due to bug in code. 6. Data corruption in memory due to hardware issue. 7. Manual modification of source data after server startup. 8. Manual modification of checksums stored in ZooKeeper. 9. Part format related settings like 'enable_mixed_granularity_parts' are different on different replicas. We will download merged part from replica to force byte-identical result. List of files in local parts: Executing query SELECT merge_algorithm, part_type FROM system.part_log WHERE event_type = 'MergeParts' AND table = 't_vertical_merges' AND part_name = 'all_0_3_3'; on node2 Executing query select 20 on node OK extractURLParameters Checking toInt16OrDefault('foo') Executing query select toInt16OrDefault('foo') on backward Executing query SELECT merge_algorithm, part_type FROM system.part_log WHERE event_type = 'MergeParts' AND table = 't_vertical_merges' AND part_name = 'all_0_3_3'; on node1 Executing query select toInt16OrDefault('foo') on upstream Command:[docker compose --env-file /ClickHouse/tests/integration/test_backward_compatibility/_instances-vertical_merges_from_compact_parts-0-gw1/.env --project-name roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1 --file /ClickHouse/tests/integration/test_backward_compatibility/_instances-vertical_merges_from_compact_parts-0-gw1/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_backward_compatibility/_instances-vertical_merges_from_compact_parts-0-gw1/node2/docker-compose.yml stop --timeout 20] [gw1] PASSED test_backward_compatibility/test_vertical_merges_from_compact_parts.py::test_vertical_merges_from_compact_parts OK toInt16OrDefault Checking sipHash128Keyed('foo') Executing query select sipHash128Keyed('foo') on backward Skipping sipHash128Keyed Checking addMicroseconds('foo') Executing query select addMicroseconds('foo') on backward Skipping addMicroseconds Checking formatDateTime('foo') Executing query select formatDateTime('foo') on backward Executing query select 20 on node Skipping formatDateTime Checking toInt8OrDefault('foo') Executing query select toInt8OrDefault('foo') on backward Executing query select toInt8OrDefault('foo') on upstream OK toInt8OrDefault Checking normalizeUTF8NFD('foo') Executing query select normalizeUTF8NFD('foo') on backward Executing query select normalizeUTF8NFD('foo') on upstream run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] OK normalizeUTF8NFD Checking leftPadUTF8('foo') Executing query select leftPadUTF8('foo') on backward Current start attempt failed. Will kill 3553 just in case. run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:True cmd: ['bash', '-c', 'kill -9 3553'] Command:[docker exec -u root roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c kill -9 3553] Stderr:bash: line 0: kill: (3553) - No such process Exitcode:1 Skipping leftPadUTF8 Checking toUInt8OrDefault('foo') Executing query select toUInt8OrDefault('foo') on backward Executing query select toUInt8OrDefault('foo') on upstream OK toUInt8OrDefault Checking s2RectIntersection('foo') Executing query select s2RectIntersection('foo') on backward Stderr: Container roottestbackuprestoreoncluster-gw0-node1-1 Stopping Stderr: Container roottestbackuprestoreoncluster-gw0-node2-1 Stopping Stderr: Container roottestbackuprestoreoncluster-gw0-node3-1 Stopping Stderr: Container roottestbackuprestoreoncluster-gw0-node2-1 Stopped Stderr: Container roottestbackuprestoreoncluster-gw0-node1-1 Stopped Stderr: Container roottestbackuprestoreoncluster-gw0-node3-1 Stopped Stderr: Container roottestbackuprestoreoncluster-gw0-zoo1-1 Stopping Stderr: Container roottestbackuprestoreoncluster-gw0-zoo2-1 Stopping Stderr: Container roottestbackuprestoreoncluster-gw0-zoo3-1 Stopping Stderr: Container roottestbackuprestoreoncluster-gw0-zoo3-1 Stopped Stderr: Container roottestbackuprestoreoncluster-gw0-zoo2-1 Stopped Stderr: Container roottestbackuprestoreoncluster-gw0-zoo1-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/node1/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/node1/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/node2/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/node2/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/node3/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/node3/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/.env --project-name roottestbackuprestoreoncluster-gw0 --file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/node2/docker-compose.yml --file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw0/node3/docker-compose.yml down --volumes] Skipping s2RectIntersection Checking mapContainsKeyLike('foo') Executing query select mapContainsKeyLike('foo') on backward Skipping mapContainsKeyLike Checking regexpQuoteMeta('foo') Executing query select regexpQuoteMeta('foo') on backward run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] No clickhouse process running. Start new one. http://localhost:None "POST /v1.46/containers/roottestbackwardcompatibilityconvertordinary-gw5-node-1/exec HTTP/1.1" 201 74 Executing query select regexpQuoteMeta('foo') on upstream http://localhost:None "POST /v1.46/exec/727a8c9b432a2f2866560d575fc194f02e4d82a5ef2dcbba059818d4d8b23682/start HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/exec/727a8c9b432a2f2866560d575fc194f02e4d82a5ef2dcbba059818d4d8b23682/json HTTP/1.1" 200 587 Stderr: Container roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node1-1 Stopping Stderr: Container roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node2-1 Stopping Stderr: Container roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node1-1 Stopped Stderr: Container roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node2-1 Stopped Stderr: Container roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-zoo3-1 Stopping Stderr: Container roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-zoo1-1 Stopping Stderr: Container roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-zoo2-1 Stopping Stderr: Container roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-zoo2-1 Stopped Stderr: Container roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-zoo1-1 Stopped Stderr: Container roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-zoo3-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_backward_compatibility/_instances-vertical_merges_from_compact_parts-0-gw1/node1/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_backward_compatibility/_instances-vertical_merges_from_compact_parts-0-gw1/node1/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_backward_compatibility/_instances-vertical_merges_from_compact_parts-0-gw1/node2/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_backward_compatibility/_instances-vertical_merges_from_compact_parts-0-gw1/node2/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_backward_compatibility/_instances-vertical_merges_from_compact_parts-0-gw1/.env --project-name roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1 --file /ClickHouse/tests/integration/test_backward_compatibility/_instances-vertical_merges_from_compact_parts-0-gw1/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_backward_compatibility/_instances-vertical_merges_from_compact_parts-0-gw1/node2/docker-compose.yml down --volumes] OK regexpQuoteMeta Checking notNullInIgnoreSet('foo') Executing query select notNullInIgnoreSet('foo') on backward Skipping notNullInIgnoreSet Checking toIPv4('foo') Executing query select toIPv4('foo') on backward Skipping toIPv4 Checking mapValues('foo') Executing query select mapValues('foo') on backward Skipping mapValues Checking toTypeName('foo') Executing query select toTypeName('foo') on backward Executing query select toTypeName('foo') on upstream OK toTypeName Checking polygonsUnionCartesian('foo') Executing query select polygonsUnionCartesian('foo') on backward Stderr: Container roottestbackuprestoreoncluster-gw0-node3-1 Stopping Stderr: Container roottestbackuprestoreoncluster-gw0-node1-1 Stopping Stderr: Container roottestbackuprestoreoncluster-gw0-node2-1 Stopping Stderr: Container roottestbackuprestoreoncluster-gw0-node3-1 Stopped Stderr: Container roottestbackuprestoreoncluster-gw0-node3-1 Removing Stderr: Container roottestbackuprestoreoncluster-gw0-node1-1 Stopped Stderr: Container roottestbackuprestoreoncluster-gw0-node1-1 Removing Stderr: Container roottestbackuprestoreoncluster-gw0-node2-1 Stopped Stderr: Container roottestbackuprestoreoncluster-gw0-node2-1 Removing Stderr: Container roottestbackuprestoreoncluster-gw0-node3-1 Removed Stderr: Container roottestbackuprestoreoncluster-gw0-node1-1 Removed Stderr: Container roottestbackuprestoreoncluster-gw0-node2-1 Removed Stderr: Container roottestbackuprestoreoncluster-gw0-zoo3-1 Stopping Stderr: Container roottestbackuprestoreoncluster-gw0-zoo1-1 Stopping Stderr: Container roottestbackuprestoreoncluster-gw0-zoo2-1 Stopping Stderr: Container roottestbackuprestoreoncluster-gw0-zoo3-1 Stopped Stderr: Container roottestbackuprestoreoncluster-gw0-zoo3-1 Removing Stderr: Container roottestbackuprestoreoncluster-gw0-zoo1-1 Stopped Stderr: Container roottestbackuprestoreoncluster-gw0-zoo1-1 Removing Stderr: Container roottestbackuprestoreoncluster-gw0-zoo2-1 Stopped Stderr: Container roottestbackuprestoreoncluster-gw0-zoo2-1 Removing Stderr: Container roottestbackuprestoreoncluster-gw0-zoo1-1 Removed Stderr: Container roottestbackuprestoreoncluster-gw0-zoo2-1 Removed Stderr: Container roottestbackuprestoreoncluster-gw0-zoo3-1 Removed Stderr: Network roottestbackuprestoreoncluster-gw0_default Removing Stderr: Network roottestbackuprestoreoncluster-gw0_default Removed Cleanup called Docker networks for project roottestbackuprestoreoncluster-gw0 are NETWORK ID NAME DRIVER SCOPE Skipping polygonsUnionCartesian Checking char('foo') Executing query select char('foo') on backward Docker containers for project roottestbackuprestoreoncluster-gw0 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestbackuprestoreoncluster-gw0 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestbackuprestoreoncluster-gw0-.*-1$' --format '{{.ID}}:{{.Names}}'] Stderr: node2 Pulling Stderr: node1 Pulling Stderr: node1 Pulled Stderr: node2 Pulled ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_backward_compatibility/_instances-normalized_count_comparison-0-gw2/.env --project-name roottestbackwardcompatibilitynormalizedcountcomparison-gw2 --file /ClickHouse/tests/integration/test_backward_compatibility/_instances-normalized_count_comparison-0-gw2/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_backward_compatibility/_instances-normalized_count_comparison-0-gw2/node2/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_backward_compatibility/_instances-normalized_count_comparison-0-gw2/.env --project-name roottestbackwardcompatibilitynormalizedcountcomparison-gw2 --file /ClickHouse/tests/integration/test_backward_compatibility/_instances-normalized_count_comparison-0-gw2/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_backward_compatibility/_instances-normalized_count_comparison-0-gw2/node2/docker-compose.yml up -d --no-recreate] Unstopped containers: {} No running containers for project: roottestbackuprestoreoncluster-gw0 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Skipping char Checking mapContains('foo') Executing query select mapContains('foo') on backward Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:1 Volumes pruned: 1 Skipping mapContains Checking toDateTime64OrNull('foo') Executing query select toDateTime64OrNull('foo') on backward 22.04: Pulling from library/ubuntu 9cb31e2e37ea: Already exists Digest: sha256:ed1544e454989078f5dec1bfdabd8c5cc9c48e0705d07b678ab6ae3fb61952d2 Status: Downloaded newer image for ubuntu:22.04 Executing query select toDateTime64OrNull('foo') on upstream OK toDateTime64OrNull Checking regionHierarchy('foo') Executing query select regionHierarchy('foo') on backward Skipping regionHierarchy Checking polygonsWithinCartesian('foo') Executing query select polygonsWithinCartesian('foo') on backward Stderr: Container roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node1-1 Stopping Stderr: Container roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node2-1 Stopping Stderr: Container roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node1-1 Stopped Stderr: Container roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node1-1 Removing Stderr: Container roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node2-1 Stopped Stderr: Container roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node2-1 Removing Stderr: Container roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node2-1 Removed Skipping polygonsWithinCartesian Stderr: Container roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-node1-1 Removed Stderr: Container roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-zoo3-1 Stopping Checking materialize('foo') Stderr: Container roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-zoo1-1 Stopping Executing query select materialize('foo') on backward Stderr: Container roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-zoo2-1 Stopping Stderr: Container roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-zoo3-1 Stopped Stderr: Container roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-zoo3-1 Removing Stderr: Container roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-zoo2-1 Stopped Stderr: Container roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-zoo2-1 Removing Stderr: Container roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-zoo1-1 Stopped Stderr: Container roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-zoo1-1 Removing Stderr: Container roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-zoo3-1 Removed Stderr: Container roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-zoo1-1 Removed Stderr: Container roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-zoo2-1 Removed Stderr: Network roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1_default Removing Stderr: Network roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1_default Removed Cleanup called Docker networks for project roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1-.*-1$' --format '{{.ID}}:{{.Names}}'] Executing query select materialize('foo') on upstream Unstopped containers: {} No running containers for project: roottestbackwardcompatibilityverticalmergesfromcompactparts-gw1 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] OK materialize Checking upperUTF8('foo') Executing query select upperUTF8('foo') on backward Stdout:1 Volumes pruned: 1 test_cancel_freeze/test.py::test_cancel_backup Running tests in /ClickHouse/tests/integration/test_cancel_freeze/test.py Cluster start called. is_up=False Stdout:4315 Clickhouse process running. run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Docker networks for project roottestcancelfreeze-gw1 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestcancelfreeze-gw1 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestcancelfreeze-gw1 are DRIVER VOLUME NAME Cleanup called Stdout:4315 Executing query select 20 on node Docker networks for project roottestcancelfreeze-gw1 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestcancelfreeze-gw1 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestcancelfreeze-gw1 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestcancelfreeze-gw1-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestcancelfreeze-gw1 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Command: docker run --rm --cpus 2 --volume /clickhouse:/usr/bin/clickhouse ubuntu:22.04 sh -c clickhouse local -q "select value from system.settings where name='max_threads'" Executing query select upperUTF8('foo') on upstream Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:1 Volumes pruned: 1 Setup directory for instance: node Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files [] to /ClickHouse/tests/integration/test_cancel_freeze/_instances-0-gw1/node/configs/config.d Setup database dir /ClickHouse/tests/integration/test_cancel_freeze/_instances-0-gw1/node/database Setup logs dir /ClickHouse/tests/integration/test_cancel_freeze/_instances-0-gw1/node/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw'} stored in /ClickHouse/tests/integration/test_cancel_freeze/_instances-0-gw1/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_cancel_freeze/_instances-0-gw1/.env --project-name roottestcancelfreeze-gw1 --file /ClickHouse/tests/integration/test_cancel_freeze/_instances-0-gw1/node/docker-compose.yml pull] OK upperUTF8 Checking multiSearchAnyCaseInsensitiveUTF8('foo') Executing query select multiSearchAnyCaseInsensitiveUTF8('foo') on backward Skipping multiSearchAnyCaseInsensitiveUTF8 Checking minSampleSizeContinous('foo') Executing query select minSampleSizeContinous('foo') on backward Stderr: Network roottestbackwardcompatibilitynormalizedcountcomparison-gw2_default Creating Stderr: Network roottestbackwardcompatibilitynormalizedcountcomparison-gw2_default Created Stderr: Container roottestbackwardcompatibilitynormalizedcountcomparison-gw2-node1-1 Creating Stderr: Container roottestbackwardcompatibilitynormalizedcountcomparison-gw2-node2-1 Creating Stderr: Container roottestbackwardcompatibilitynormalizedcountcomparison-gw2-node1-1 Created Stderr: Container roottestbackwardcompatibilitynormalizedcountcomparison-gw2-node2-1 Created Stderr: Container roottestbackwardcompatibilitynormalizedcountcomparison-gw2-node1-1 Starting Stderr: Container roottestbackwardcompatibilitynormalizedcountcomparison-gw2-node2-1 Starting Stderr: Container roottestbackwardcompatibilitynormalizedcountcomparison-gw2-node2-1 Started Stderr: Container roottestbackwardcompatibilitynormalizedcountcomparison-gw2-node1-1 Started ClickHouse instance created get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottestbackwardcompatibilitynormalizedcountcomparison-gw2-node1-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottestbackwardcompatibilitynormalizedcountcomparison-gw2-node1-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node1, ip: 172.16.2.3... http://localhost:None "GET /v1.46/containers/roottestbackwardcompatibilitynormalizedcountcomparison-gw2-node1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/39c4f77e0aa4e735a003ed918cb6c24f77c845e3392b1ea76a38c605c55d4f0a/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/39c4f77e0aa4e735a003ed918cb6c24f77c845e3392b1ea76a38c605c55d4f0a/json HTTP/1.1" 200 None Skipping minSampleSizeContinous Checking toYYYYMM('foo') Executing query select toYYYYMM('foo') on backward http://localhost:None "GET /v1.46/containers/39c4f77e0aa4e735a003ed918cb6c24f77c845e3392b1ea76a38c605c55d4f0a/json HTTP/1.1" 200 None Skipping toYYYYMM Checking domainWithoutWWWRFC('foo') Executing query select domainWithoutWWWRFC('foo') on backward Executing query select 20 on node http://localhost:None "GET /v1.46/containers/39c4f77e0aa4e735a003ed918cb6c24f77c845e3392b1ea76a38c605c55d4f0a/json HTTP/1.1" 200 None Executing query select domainWithoutWWWRFC('foo') on upstream Command: docker run --rm --cpus 4 --volume /clickhouse:/usr/bin/clickhouse ubuntu:22.04 sh -c clickhouse local -q "select value from system.settings where name='max_threads'" http://localhost:None "GET /v1.46/containers/39c4f77e0aa4e735a003ed918cb6c24f77c845e3392b1ea76a38c605c55d4f0a/json HTTP/1.1" 200 None ClickHouse node1 started get_instance_ip instance_name=node2 http://localhost:None "GET /v1.46/containers/roottestbackwardcompatibilitynormalizedcountcomparison-gw2-node2-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node2 http://localhost:None "GET /v1.46/containers/roottestbackwardcompatibilitynormalizedcountcomparison-gw2-node2-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node2, ip: 172.16.2.2... http://localhost:None "GET /v1.46/containers/roottestbackwardcompatibilitynormalizedcountcomparison-gw2-node2-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/f750dfa427a48b834d59a91ef7f3321973067112a655a761d95bde7287633469/json HTTP/1.1" 200 None ClickHouse node2 started Executing query create table tab (x UInt64, y String, z Nullable(Int64)) engine = Memory on node1 OK domainWithoutWWWRFC Checking arrayEnumerateDenseRanked('foo') Executing query select arrayEnumerateDenseRanked('foo') on backward Skipping arrayEnumerateDenseRanked Checking queryString('foo') Executing query select queryString('foo') on backward Executing query select queryString('foo') on upstream Executing query create table tab (x UInt64, y String, z Nullable(Int64)) engine = Memory on node2 OK queryString Checking getSubcolumn('foo') Executing query select getSubcolumn('foo') on backward Executing query insert into tab values (1, 'a', null) on node1 Skipping getSubcolumn Checking reinterpretAsUUID('foo') Executing query select reinterpretAsUUID('foo') on backward Executing query insert into tab values (1, 'a', null) on node2 Executing query select 20 on node Command: docker run --rm --cpus 2.8 --volume /clickhouse:/usr/bin/clickhouse ubuntu:22.04 sh -c clickhouse local -q "select value from system.settings where name='max_threads'" Executing query select reinterpretAsUUID('foo') on upstream Executing query select count(), count(1), count(x), count(y), count(z) from remote('node{1,2}', default, tab) on node1 OK reinterpretAsUUID Checking simpleJSONExtractFloat('foo') Executing query select simpleJSONExtractFloat('foo') on backward Executing query select count(), count(1), count(x), count(y), count(z) from remote('node{1,2}', default, tab) on node2 Skipping simpleJSONExtractFloat Checking h3HexAreaM2('foo') Executing query select h3HexAreaM2('foo') on backward Skipping h3HexAreaM2 Checking toYYYYMMDDhhmmss('foo') Executing query select toYYYYMMDDhhmmss('foo') on backward Executing query drop table tab on node1 Skipping toYYYYMMDDhhmmss Checking h3NumHexagons('foo') Executing query select h3NumHexagons('foo') on backward Executing query drop table tab on node2 Skipping h3NumHexagons Checking parseDateTime64BestEffortUSOrNull('foo') Executing query select parseDateTime64BestEffortUSOrNull('foo') on backward Executing query select 20 on node Command:[docker compose --env-file /ClickHouse/tests/integration/test_backward_compatibility/_instances-normalized_count_comparison-0-gw2/.env --project-name roottestbackwardcompatibilitynormalizedcountcomparison-gw2 --file /ClickHouse/tests/integration/test_backward_compatibility/_instances-normalized_count_comparison-0-gw2/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_backward_compatibility/_instances-normalized_count_comparison-0-gw2/node2/docker-compose.yml stop --timeout 20] [gw2] PASSED test_backward_compatibility/test_normalized_count_comparison.py::test_select_aggregate_alias_column [gw8] PASSED test_cgroup_limit/test.py::test_cgroup_cpu_limit Executing query select parseDateTime64BestEffortUSOrNull('foo') on upstream OK parseDateTime64BestEffortUSOrNull Checking moduloLegacy('foo') Executing query select moduloLegacy('foo') on backward Skipping moduloLegacy Checking dictGetInt32('foo') Executing query select dictGetInt32('foo') on backward Skipping dictGetInt32 Checking mortonEncode('foo') Executing query select mortonEncode('foo') on backward Skipping mortonEncode Checking mapExtractKeyLike('foo') Executing query select mapExtractKeyLike('foo') on backward Skipping mapExtractKeyLike Checking IPv6StringToNumOrNull('foo') Executing query select IPv6StringToNumOrNull('foo') on backward Executing query select IPv6StringToNumOrNull('foo') on upstream Executing query select 20 on node OK IPv6StringToNumOrNull Checking arrayResize('foo') Executing query select arrayResize('foo') on backward Skipping arrayResize Checking randBernoulli('foo') Executing query select randBernoulli('foo') on backward Skipping randBernoulli Checking addWeeks('foo') Executing query select addWeeks('foo') on backward Skipping addWeeks Checking rightPadUTF8('foo') Executing query select rightPadUTF8('foo') on backward Skipping rightPadUTF8 Checking uniqThetaIntersect('foo') Executing query select uniqThetaIntersect('foo') on backward Skipping uniqThetaIntersect Checking cosh('foo') Executing query select cosh('foo') on backward Skipping cosh Checking __bitBoolMaskOr('foo') Executing query select __bitBoolMaskOr('foo') on backward Skipping __bitBoolMaskOr Checking geohashEncode('foo') Executing query select geohashEncode('foo') on backward Skipping geohashEncode Checking formatReadableTimeDelta('foo') Executing query select formatReadableTimeDelta('foo') on backward Executing query select 20 on node Skipping formatReadableTimeDelta Checking randExponential('foo') Executing query select randExponential('foo') on backward Skipping randExponential Checking randomString('foo') Executing query select randomString('foo') on backward Skipping randomString Checking normalizeUTF8NFKD('foo') Executing query select normalizeUTF8NFKD('foo') on backward Executing query select normalizeUTF8NFKD('foo') on upstream OK normalizeUTF8NFKD Checking sigmoid('foo') Executing query select sigmoid('foo') on backward Skipping sigmoid Checking normalizeUTF8NFC('foo') Executing query select normalizeUTF8NFC('foo') on backward Executing query select normalizeUTF8NFC('foo') on upstream OK normalizeUTF8NFC Checking multiSearchAllPositions('foo') Executing query select multiSearchAllPositions('foo') on backward Executing query select 20 on node Skipping multiSearchAllPositions Checking subBitmap('foo') Executing query select subBitmap('foo') on backward Skipping subBitmap Checking defaultValueOfArgumentType('foo') Executing query select defaultValueOfArgumentType('foo') on backward Connection dropped: socket connection error: Connection refused Connection dropped: socket connection error: Connection refused Executing query select defaultValueOfArgumentType('foo') on upstream OK defaultValueOfArgumentType Checking mortonDecode('foo') Executing query select mortonDecode('foo') on backward Skipping mortonDecode Checking roundDuration('foo') Executing query select roundDuration('foo') on backward Skipping roundDuration Checking multiSearchFirstIndexCaseInsensitiveUTF8('foo') Executing query select multiSearchFirstIndexCaseInsensitiveUTF8('foo') on backward Skipping multiSearchFirstIndexCaseInsensitiveUTF8 Checking positionCaseInsensitiveUTF8('foo') Executing query select positionCaseInsensitiveUTF8('foo') on backward Skipping positionCaseInsensitiveUTF8 Checking parseDateTime32BestEffortOrZero('foo') Executing query select parseDateTime32BestEffortOrZero('foo') on backward Executing query select 20 on node Executing query select parseDateTime32BestEffortOrZero('foo') on upstream OK parseDateTime32BestEffortOrZero Checking blockNumber('foo') Executing query select blockNumber('foo') on backward Skipping blockNumber Checking rightUTF8('foo') Executing query select rightUTF8('foo') on backward Skipping rightUTF8 Checking cutURLParameter('foo') Executing query select cutURLParameter('foo') on backward Skipping cutURLParameter Checking max2('foo') Executing query select max2('foo') on backward Connection dropped: socket connection error: Connection refused Skipping max2 Checking multiSearchFirstPositionCaseInsensitive('foo') Executing query select multiSearchFirstPositionCaseInsensitive('foo') on backward Skipping multiSearchFirstPositionCaseInsensitive Checking indexOf('foo') Executing query select indexOf('foo') on backward Skipping indexOf Checking uptime('foo') Executing query select uptime('foo') on backward Connection dropped: socket connection error: Connection refused Skipping uptime Checking toModifiedJulianDayOrNull('foo') Executing query select toModifiedJulianDayOrNull('foo') on backward Executing query select 20 on node Executing query select toModifiedJulianDayOrNull('foo') on upstream OK toModifiedJulianDayOrNull Checking multiplyDecimal('foo') Executing query select multiplyDecimal('foo') on backward Skipping multiplyDecimal Checking alphaTokens('foo') Executing query select alphaTokens('foo') on backward Executing query select alphaTokens('foo') on upstream OK alphaTokens Checking arrayFirstIndex('foo') Executing query select arrayFirstIndex('foo') on backward Skipping arrayFirstIndex Checking arrayPartialShuffle('foo') Executing query select arrayPartialShuffle('foo') on backward Skipping arrayPartialShuffle Checking negate('foo') Executing query select negate('foo') on backward Executing query select 20 on node Skipping negate Checking joinGet('foo') Executing query select joinGet('foo') on backward Stderr: Container roottestbackwardcompatibilitynormalizedcountcomparison-gw2-node2-1 Stopping Stderr: Container roottestbackwardcompatibilitynormalizedcountcomparison-gw2-node1-1 Stopping Stderr: Container roottestbackwardcompatibilitynormalizedcountcomparison-gw2-node2-1 Stopped Stderr: Container roottestbackwardcompatibilitynormalizedcountcomparison-gw2-node1-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_backward_compatibility/_instances-normalized_count_comparison-0-gw2/node1/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_backward_compatibility/_instances-normalized_count_comparison-0-gw2/node1/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_backward_compatibility/_instances-normalized_count_comparison-0-gw2/node2/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_backward_compatibility/_instances-normalized_count_comparison-0-gw2/node2/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Skipping joinGet Checking geoToH3('foo') Executing query select geoToH3('foo') on backward Command:[docker compose --env-file /ClickHouse/tests/integration/test_backward_compatibility/_instances-normalized_count_comparison-0-gw2/.env --project-name roottestbackwardcompatibilitynormalizedcountcomparison-gw2 --file /ClickHouse/tests/integration/test_backward_compatibility/_instances-normalized_count_comparison-0-gw2/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_backward_compatibility/_instances-normalized_count_comparison-0-gw2/node2/docker-compose.yml down --volumes] Skipping geoToH3 Checking firstSignificantSubdomainCustom('foo') Executing query select firstSignificantSubdomainCustom('foo') on backward Skipping firstSignificantSubdomainCustom Checking toUnixTimestamp64Milli('foo') Executing query select toUnixTimestamp64Milli('foo') on backward Skipping toUnixTimestamp64Milli Checking simpleJSONExtractBool('foo') Executing query select simpleJSONExtractBool('foo') on backward Skipping simpleJSONExtractBool Checking normalizeUTF8NFKC('foo') Executing query select normalizeUTF8NFKC('foo') on backward Executing query select normalizeUTF8NFKC('foo') on upstream OK normalizeUTF8NFKC Checking parseDateTimeBestEffortUSOrNull('foo') Executing query select parseDateTimeBestEffortUSOrNull('foo') on backward run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stderr: Container roottestbackwardcompatibilitynormalizedcountcomparison-gw2-node2-1 Stopping Stderr: Container roottestbackwardcompatibilitynormalizedcountcomparison-gw2-node1-1 Stopping Stderr: Container roottestbackwardcompatibilitynormalizedcountcomparison-gw2-node2-1 Stopped Stderr: Container roottestbackwardcompatibilitynormalizedcountcomparison-gw2-node2-1 Removing Stderr: Container roottestbackwardcompatibilitynormalizedcountcomparison-gw2-node1-1 Stopped Stderr: Container roottestbackwardcompatibilitynormalizedcountcomparison-gw2-node1-1 Removing Stderr: Container roottestbackwardcompatibilitynormalizedcountcomparison-gw2-node2-1 Removed Stderr: Container roottestbackwardcompatibilitynormalizedcountcomparison-gw2-node1-1 Removed Stderr: Network roottestbackwardcompatibilitynormalizedcountcomparison-gw2_default Removing Stderr: Network roottestbackwardcompatibilitynormalizedcountcomparison-gw2_default Removed Cleanup called Docker networks for project roottestbackwardcompatibilitynormalizedcountcomparison-gw2 are NETWORK ID NAME DRIVER SCOPE Executing query select parseDateTimeBestEffortUSOrNull('foo') on upstream Docker containers for project roottestbackwardcompatibilitynormalizedcountcomparison-gw2 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Current start attempt failed. Will kill 4315 just in case. run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:True cmd: ['bash', '-c', 'kill -9 4315'] Command:[docker exec -u root roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c kill -9 4315] Docker volumes for project roottestbackwardcompatibilitynormalizedcountcomparison-gw2 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestbackwardcompatibilitynormalizedcountcomparison-gw2-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestbackwardcompatibilitynormalizedcountcomparison-gw2 Trying to prune unused networks... Stderr:bash: line 0: kill: (4315) - No such process Exitcode:1 Trying to prune unused images... Command:[docker image prune -f] OK parseDateTimeBestEffortUSOrNull Checking detectProgrammingLanguage('foo') Executing query select detectProgrammingLanguage('foo') on backward Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:1 Volumes pruned: 1 Skipping detectProgrammingLanguage Checking neighbor('foo') Executing query select neighbor('foo') on backward Skipping neighbor Checking toUInt256OrZero('foo') Executing query select toUInt256OrZero('foo') on backward Executing query select toUInt256OrZero('foo') on upstream OK toUInt256OrZero Checking toStringCutToZero('foo') Executing query select toStringCutToZero('foo') on backward Executing query select toStringCutToZero('foo') on upstream OK toStringCutToZero Checking toUInt8OrZero('foo') Executing query select toUInt8OrZero('foo') on backward Executing query select toUInt8OrZero('foo') on upstream OK toUInt8OrZero Checking not('foo') Executing query select not('foo') on backward run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Skipping not Checking reinterpretAsInt8('foo') Executing query select reinterpretAsInt8('foo') on backward No clickhouse process running. Start new one. http://localhost:None "POST /v1.46/containers/roottestbackwardcompatibilityconvertordinary-gw5-node-1/exec HTTP/1.1" 201 74 http://localhost:None "POST /v1.46/exec/c1f5693924e7a8b88db1064c8177ad2a32cdeb213c3a5b7257295a7a518984fa/start HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/exec/c1f5693924e7a8b88db1064c8177ad2a32cdeb213c3a5b7257295a7a518984fa/json HTTP/1.1" 200 587 Executing query select reinterpretAsInt8('foo') on upstream OK reinterpretAsInt8 Checking normalizedQueryHashKeepNames('foo') Executing query select normalizedQueryHashKeepNames('foo') on backward Executing query select normalizedQueryHashKeepNames('foo') on upstream OK normalizedQueryHashKeepNames Checking tryDecrypt('foo') Executing query select tryDecrypt('foo') on backward Skipping tryDecrypt Checking log10('foo') Executing query select log10('foo') on backward Skipping log10 Checking globalIn('foo') Executing query select globalIn('foo') on backward Skipping globalIn Checking JSONExtractArrayRaw('foo') Executing query select JSONExtractArrayRaw('foo') on backward Executing query select JSONExtractArrayRaw('foo') on upstream OK JSONExtractArrayRaw Checking version('foo') Executing query select version('foo') on backward Skipping version Checking joinGetOrNull('foo') Executing query select joinGetOrNull('foo') on backward Skipping joinGetOrNull Checking notILike('foo') Executing query select notILike('foo') on backward Skipping notILike Checking geohashDecode('foo') Executing query select geohashDecode('foo') on backward Executing query select geohashDecode('foo') on upstream OK geohashDecode Checking toModifiedJulianDay('foo') Executing query select toModifiedJulianDay('foo') on backward Skipping toModifiedJulianDay Checking notLike('foo') Executing query select notLike('foo') on backward Skipping notLike Checking fragment('foo') Executing query select fragment('foo') on backward Executing query select fragment('foo') on upstream OK fragment Checking match('foo') Executing query select match('foo') on backward run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:5072 Clickhouse process running. run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Skipping match Checking greater('foo') Executing query select greater('foo') on backward Stdout:5072 Executing query select 20 on node Skipping greater Checking toRelativeMonthNum('foo') Executing query select toRelativeMonthNum('foo') on backward Skipping toRelativeMonthNum Checking emptyArrayString('foo') Executing query select emptyArrayString('foo') on backward Skipping emptyArrayString Checking bitOr('foo') Executing query select bitOr('foo') on backward Skipping bitOr Checking sleep('foo') Executing query select sleep('foo') on backward Skipping sleep Checking arrayElement('foo') Executing query select arrayElement('foo') on backward Skipping arrayElement Checking tgamma('foo') Executing query select tgamma('foo') on backward Skipping tgamma Checking emptyArrayToSingle('foo') Executing query select emptyArrayToSingle('foo') on backward Skipping emptyArrayToSingle Checking reinterpretAsInt128('foo') Executing query select reinterpretAsInt128('foo') on backward Executing query select 20 on node Executing query select reinterpretAsInt128('foo') on upstream OK reinterpretAsInt128 Checking h3ToChildren('foo') Executing query select h3ToChildren('foo') on backward Skipping h3ToChildren Checking subtractNanoseconds('foo') Executing query select subtractNanoseconds('foo') on backward Skipping subtractNanoseconds Checking queryStringAndFragment('foo') Executing query select queryStringAndFragment('foo') on backward Executing query select queryStringAndFragment('foo') on upstream OK queryStringAndFragment Checking toUInt32OrDefault('foo') Executing query select toUInt32OrDefault('foo') on backward Executing query select toUInt32OrDefault('foo') on upstream OK toUInt32OrDefault Checking emptyArrayFloat32('foo') Executing query select emptyArrayFloat32('foo') on backward Executing query select 20 on node Skipping emptyArrayFloat32 Checking MACStringToOUI('foo') Executing query select MACStringToOUI('foo') on backward Executing query select MACStringToOUI('foo') on upstream OK MACStringToOUI Checking file('foo') Executing query select file('foo') on backward Skipping file Checking roundAge('foo') Executing query select roundAge('foo') on backward Skipping roundAge Checking simpleJSONExtractString('foo') Executing query select simpleJSONExtractString('foo') on backward Skipping simpleJSONExtractString Checking toUInt128('foo') Executing query select toUInt128('foo') on backward Skipping toUInt128 Checking emptyArrayUInt64('foo') Executing query select emptyArrayUInt64('foo') on backward Skipping emptyArrayUInt64 Checking h3EdgeAngle('foo') Executing query select h3EdgeAngle('foo') on backward Skipping h3EdgeAngle Checking polygonPerimeterCartesian('foo') Executing query select polygonPerimeterCartesian('foo') on backward Executing query select 20 on node Skipping polygonPerimeterCartesian Checking exp2('foo') Executing query select exp2('foo') on backward Skipping exp2 Checking concatAssumeInjective('foo') Executing query select concatAssumeInjective('foo') on backward Skipping concatAssumeInjective Checking h3IsValid('foo') Executing query select h3IsValid('foo') on backward Skipping h3IsValid Checking emptyArrayUInt32('foo') Executing query select emptyArrayUInt32('foo') on backward Skipping emptyArrayUInt32 Checking dotProduct('foo') Executing query select dotProduct('foo') on backward Skipping dotProduct Checking emptyArrayUInt16('foo') Executing query select emptyArrayUInt16('foo') on backward Skipping emptyArrayUInt16 Checking toDateTimeOrZero('foo') Executing query select toDateTimeOrZero('foo') on backward Executing query select toDateTimeOrZero('foo') on upstream Executing query select 20 on node OK toDateTimeOrZero Checking nullIf('foo') Executing query select nullIf('foo') on backward Skipping nullIf Checking toMonday('foo') Executing query select toMonday('foo') on backward Skipping toMonday Checking dictGetInt32OrDefault('foo') Executing query select dictGetInt32OrDefault('foo') on backward Skipping dictGetInt32OrDefault Checking dictGetIPv6OrDefault('foo') Executing query select dictGetIPv6OrDefault('foo') on backward Skipping dictGetIPv6OrDefault Checking toYear('foo') Executing query select toYear('foo') on backward Skipping toYear Checking toDateTime64OrZero('foo') Executing query select toDateTime64OrZero('foo') on backward Executing query select toDateTime64OrZero('foo') on upstream Executing query select 20 on node OK toDateTime64OrZero Checking subtractSeconds('foo') Executing query select subtractSeconds('foo') on backward Skipping subtractSeconds Checking tid('foo') Executing query select tid('foo') on backward Skipping tid Checking countSubstrings('foo') Executing query select countSubstrings('foo') on backward Skipping countSubstrings Checking reinterpretAsFloat64('foo') Executing query select reinterpretAsFloat64('foo') on backward Executing query select reinterpretAsFloat64('foo') on upstream OK reinterpretAsFloat64 Checking tupleMultiplyByNumber('foo') Executing query select tupleMultiplyByNumber('foo') on backward Skipping tupleMultiplyByNumber Checking mapFromArrays('foo') Executing query select mapFromArrays('foo') on backward Skipping mapFromArrays Checking cutToFirstSignificantSubdomainCustomWithWWWRFC('foo') Executing query select cutToFirstSignificantSubdomainCustomWithWWWRFC('foo') on backward Skipping cutToFirstSignificantSubdomainCustomWithWWWRFC Checking multiSearchAllPositionsCaseInsensitive('foo') Executing query select multiSearchAllPositionsCaseInsensitive('foo') on backward Skipping multiSearchAllPositionsCaseInsensitive Checking DATE('foo') Executing query select DATE('foo') on backward Skipping DATE Checking h3CellAreaRads2('foo') Executing query select h3CellAreaRads2('foo') on backward Executing query select 20 on node Skipping h3CellAreaRads2 Checking pointInEllipses('foo') Executing query select pointInEllipses('foo') on backward Skipping pointInEllipses Checking splitByWhitespace('foo') Executing query select splitByWhitespace('foo') on backward Executing query select splitByWhitespace('foo') on upstream OK splitByWhitespace Checking toInt256OrZero('foo') Executing query select toInt256OrZero('foo') on backward Stderr: node Pulling Stderr: node Pulled ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_cancel_freeze/_instances-0-gw1/.env --project-name roottestcancelfreeze-gw1 --file /ClickHouse/tests/integration/test_cancel_freeze/_instances-0-gw1/node/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_cancel_freeze/_instances-0-gw1/.env --project-name roottestcancelfreeze-gw1 --file /ClickHouse/tests/integration/test_cancel_freeze/_instances-0-gw1/node/docker-compose.yml up -d --no-recreate] Executing query select toInt256OrZero('foo') on upstream OK toInt256OrZero Checking pointInPolygon('foo') Executing query select pointInPolygon('foo') on backward Skipping pointInPolygon Checking polygonsIntersectionCartesian('foo') Executing query select polygonsIntersectionCartesian('foo') on backward Executing query select 20 on node Skipping polygonsIntersectionCartesian Checking polygonPerimeterSpherical('foo') Executing query select polygonPerimeterSpherical('foo') on backward Skipping polygonPerimeterSpherical Checking e('foo') Executing query select e('foo') on backward Skipping e Checking buildId('foo') Executing query select buildId('foo') on backward Skipping buildId Checking FQDN('foo') Executing query select FQDN('foo') on backward Stderr: Network roottestcancelfreeze-gw1_default Creating Stderr: Network roottestcancelfreeze-gw1_default Created Stderr: Container roottestcancelfreeze-gw1-node-1 Creating Stderr: Container roottestcancelfreeze-gw1-node-1 Created Stderr: Container roottestcancelfreeze-gw1-node-1 Starting Stderr: Container roottestcancelfreeze-gw1-node-1 Started ClickHouse instance created get_instance_ip instance_name=node http://localhost:None "GET /v1.46/containers/roottestcancelfreeze-gw1-node-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node http://localhost:None "GET /v1.46/containers/roottestcancelfreeze-gw1-node-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node, ip: 172.16.2.2... http://localhost:None "GET /v1.46/containers/roottestcancelfreeze-gw1-node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/8f635427dcfd96ff7ccb3a0080ebfb570961b282368962549ec6cf531d5259ea/json HTTP/1.1" 200 None Skipping FQDN Checking mapUpdate('foo') Executing query select mapUpdate('foo') on backward http://localhost:None "GET /v1.46/containers/8f635427dcfd96ff7ccb3a0080ebfb570961b282368962549ec6cf531d5259ea/json HTTP/1.1" 200 None Skipping mapUpdate Checking currentProfiles('foo') Executing query select currentProfiles('foo') on backward Skipping currentProfiles Checking isNull('foo') Executing query select isNull('foo') on backward http://localhost:None "GET /v1.46/containers/8f635427dcfd96ff7ccb3a0080ebfb570961b282368962549ec6cf531d5259ea/json HTTP/1.1" 200 None Executing query select isNull('foo') on upstream Executing query select 20 on node http://localhost:None "GET /v1.46/containers/8f635427dcfd96ff7ccb3a0080ebfb570961b282368962549ec6cf531d5259ea/json HTTP/1.1" 200 None OK isNull Checking isNaN('foo') Executing query select isNaN('foo') on backward Skipping isNaN Checking multiSearchAllPositionsCaseInsensitiveUTF8('foo') Executing query select multiSearchAllPositionsCaseInsensitiveUTF8('foo') on backward http://localhost:None "GET /v1.46/containers/8f635427dcfd96ff7ccb3a0080ebfb570961b282368962549ec6cf531d5259ea/json HTTP/1.1" 200 None ClickHouse node started Executing query CREATE TABLE tbl (x UInt32, y UInt32) ENGINE=MergeTree() PARTITION BY (x%20000) ORDER BY x on node Skipping multiSearchAllPositionsCaseInsensitiveUTF8 Checking subtractDays('foo') Executing query select subtractDays('foo') on backward Executing query INSERT INTO tbl SELECT number, number FROM numbers(20000) SETTINGS max_partitions_per_insert_block=20000 on node Skipping subtractDays Checking visibleWidth('foo') Executing query select visibleWidth('foo') on backward Executing query select visibleWidth('foo') on upstream OK visibleWidth Checking toStartOfWeek('foo') Executing query select toStartOfWeek('foo') on backward Skipping toStartOfWeek Checking multiFuzzyMatchAny('foo') Executing query select multiFuzzyMatchAny('foo') on backward Skipping multiFuzzyMatchAny Checking flattenTuple('foo') Executing query select flattenTuple('foo') on backward Skipping flattenTuple Checking polygonsWithinSpherical('foo') Executing query select polygonsWithinSpherical('foo') on backward Skipping polygonsWithinSpherical Checking timezoneOffset('foo') Executing query select timezoneOffset('foo') on backward Executing query select 20 on node Skipping timezoneOffset Checking transform('foo') Executing query select transform('foo') on backward Skipping transform Checking encodeXMLComponent('foo') Executing query select encodeXMLComponent('foo') on backward Executing query select encodeXMLComponent('foo') on upstream OK encodeXMLComponent Checking h3Distance('foo') Executing query select h3Distance('foo') on backward Skipping h3Distance Checking L2Normalize('foo') Executing query select L2Normalize('foo') on backward Skipping L2Normalize Checking pow('foo') Executing query select pow('foo') on backward Skipping pow Checking toFloat32OrDefault('foo') Executing query select toFloat32OrDefault('foo') on backward Executing query select toFloat32OrDefault('foo') on upstream OK toFloat32OrDefault Checking arrayDistinct('foo') Executing query select arrayDistinct('foo') on backward Skipping arrayDistinct Checking dictGetUInt8('foo') Executing query select dictGetUInt8('foo') on backward run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Current start attempt failed. Will kill 5072 just in case. run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:True cmd: ['bash', '-c', 'kill -9 5072'] Command:[docker exec -u root roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c kill -9 5072] Skipping dictGetUInt8 Checking emptyArrayUInt8('foo') Executing query select emptyArrayUInt8('foo') on backward Stderr:bash: line 0: kill: (5072) - No such process Exitcode:1 Skipping emptyArrayUInt8 Checking positionCaseInsensitive('foo') Executing query select positionCaseInsensitive('foo') on backward Skipping positionCaseInsensitive Checking acosh('foo') Executing query select acosh('foo') on backward Skipping acosh Checking toRelativeMinuteNum('foo') Executing query select toRelativeMinuteNum('foo') on backward Skipping toRelativeMinuteNum Checking IPv6NumToString('foo') Executing query select IPv6NumToString('foo') on backward Skipping IPv6NumToString Checking throwIf('foo') Executing query select throwIf('foo') on backward Skipping throwIf Checking randUniform('foo') Executing query select randUniform('foo') on backward Skipping randUniform Checking initializeAggregation('foo') Executing query select initializeAggregation('foo') on backward Skipping initializeAggregation Checking toDateTime64('foo') Executing query select toDateTime64('foo') on backward Skipping toDateTime64 Checking yesterday('foo') Executing query select yesterday('foo') on backward run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Skipping yesterday Checking toBool('foo') Executing query select toBool('foo') on backward No clickhouse process running. Start new one. http://localhost:None "POST /v1.46/containers/roottestbackwardcompatibilityconvertordinary-gw5-node-1/exec HTTP/1.1" 201 74 http://localhost:None "POST /v1.46/exec/5c65b9a5d59f2fc157f433aa450b07785f94fddbbccc498aa734eef7c729c994/start HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/exec/5c65b9a5d59f2fc157f433aa450b07785f94fddbbccc498aa734eef7c729c994/json HTTP/1.1" 200 587 Skipping toBool Checking randPoisson('foo') Executing query select randPoisson('foo') on backward Skipping randPoisson Checking nowInBlock('foo') Executing query select nowInBlock('foo') on backward Skipping nowInBlock Checking randNegativeBinomial('foo') Executing query select randNegativeBinomial('foo') on backward Skipping randNegativeBinomial Checking toUnixTimestamp64Micro('foo') Executing query select toUnixTimestamp64Micro('foo') on backward Skipping toUnixTimestamp64Micro Checking randBinomial('foo') Executing query select randBinomial('foo') on backward Skipping randBinomial Checking toInt32OrZero('foo') Executing query select toInt32OrZero('foo') on backward Executing query select toInt32OrZero('foo') on upstream OK toInt32OrZero Checking randStudentT('foo') Executing query select randStudentT('foo') on backward Skipping randStudentT Checking aes_encrypt_mysql('foo') Executing query select aes_encrypt_mysql('foo') on backward Skipping aes_encrypt_mysql Checking hasTokenOrNull('foo') Executing query select hasTokenOrNull('foo') on backward Skipping hasTokenOrNull Checking tuple('foo') Executing query select tuple('foo') on backward Executing query select tuple('foo') on upstream OK tuple Checking ngramMinHashCaseInsensitive('foo') Executing query select ngramMinHashCaseInsensitive('foo') on backward Executing query select ngramMinHashCaseInsensitive('foo') on upstream OK ngramMinHashCaseInsensitive Checking toInt64OrDefault('foo') Executing query select toInt64OrDefault('foo') on backward Executing query select toInt64OrDefault('foo') on upstream run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] OK toInt64OrDefault Stdout:5833 Clickhouse process running. Checking toISOYear('foo') run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Executing query select toISOYear('foo') on backward Command:[docker exec roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:5833 Executing query select 20 on node Skipping toISOYear Checking randFisherF('foo') Executing query select randFisherF('foo') on backward Skipping randFisherF Checking reinterpretAsFixedString('foo') Executing query select reinterpretAsFixedString('foo') on backward Skipping reinterpretAsFixedString Checking toLastDayOfMonth('foo') Executing query select toLastDayOfMonth('foo') on backward Skipping toLastDayOfMonth Checking randomPrintableASCII('foo') Executing query select randomPrintableASCII('foo') on backward Skipping randomPrintableASCII Checking snowflakeToDateTime('foo') Executing query select snowflakeToDateTime('foo') on backward Skipping snowflakeToDateTime Checking notEquals('foo') Executing query select notEquals('foo') on backward Skipping notEquals Checking reinterpretAsDateTime('foo') Executing query select reinterpretAsDateTime('foo') on backward Executing query select reinterpretAsDateTime('foo') on upstream Executing query select 20 on node OK reinterpretAsDateTime Checking multiSearchAnyCaseInsensitive('foo') Executing query select multiSearchAnyCaseInsensitive('foo') on backward Skipping multiSearchAnyCaseInsensitive Checking s2GetNeighbors('foo') Executing query select s2GetNeighbors('foo') on backward Skipping s2GetNeighbors Checking arrayAll('foo') Executing query select arrayAll('foo') on backward Skipping arrayAll Checking toIntervalDay('foo') Executing query select toIntervalDay('foo') on backward Skipping toIntervalDay Checking reinterpretAsFloat32('foo') Executing query select reinterpretAsFloat32('foo') on backward Executing query select reinterpretAsFloat32('foo') on upstream OK reinterpretAsFloat32 Checking topLevelDomain('foo') Executing query select topLevelDomain('foo') on backward Executing query select topLevelDomain('foo') on upstream OK topLevelDomain Checking greatest('foo') Executing query select greatest('foo') on backward Executing query select 20 on node Executing query select greatest('foo') on upstream Connection dropped: socket connection error: No route to host Connection dropped: socket connection error: No route to host OK greatest Checking toStartOfMillisecond('foo') Executing query select toStartOfMillisecond('foo') on backward Skipping toStartOfMillisecond Checking reinterpretAsInt64('foo') Executing query select reinterpretAsInt64('foo') on backward Executing query select reinterpretAsInt64('foo') on upstream OK reinterpretAsInt64 Checking reinterpretAsInt16('foo') Executing query select reinterpretAsInt16('foo') on backward Executing query select reinterpretAsInt16('foo') on upstream OK reinterpretAsInt16 Checking and('foo') Executing query select and('foo') on backward Executing query select 20 on node Skipping and Checking detectLanguageUnknown('foo') Executing query select detectLanguageUnknown('foo') on backward Skipping detectLanguageUnknown Checking reinterpretAsUInt256('foo') Executing query select reinterpretAsUInt256('foo') on backward Executing query select reinterpretAsUInt256('foo') on upstream OK reinterpretAsUInt256 Checking reinterpretAsUInt64('foo') Executing query select reinterpretAsUInt64('foo') on backward Executing query select reinterpretAsUInt64('foo') on upstream OK reinterpretAsUInt64 Checking subtractYears('foo') Executing query select subtractYears('foo') on backward Skipping subtractYears Checking readWKTMultiPolygon('foo') Executing query select readWKTMultiPolygon('foo') on backward Skipping readWKTMultiPolygon Checking reinterpretAsUInt16('foo') Executing query select reinterpretAsUInt16('foo') on backward Executing query select 20 on node Executing query select reinterpretAsUInt16('foo') on upstream OK reinterpretAsUInt16 Checking fromUnixTimestamp('foo') Executing query select fromUnixTimestamp('foo') on backward Skipping fromUnixTimestamp Checking hasThreadFuzzer('foo') Executing query select hasThreadFuzzer('foo') on backward Skipping hasThreadFuzzer Checking arrayAUC('foo') Executing query select arrayAUC('foo') on backward Skipping arrayAUC Checking multiMatchAllIndices('foo') Executing query select multiMatchAllIndices('foo') on backward Skipping multiMatchAllIndices Checking replaceRegexpAll('foo') Executing query select replaceRegexpAll('foo') on backward Skipping replaceRegexpAll Checking parseDateTimeInJodaSyntax('foo') Executing query select parseDateTimeInJodaSyntax('foo') on backward Skipping parseDateTimeInJodaSyntax Checking timeSlot('foo') Executing query select timeSlot('foo') on backward Executing query select 20 on node Skipping timeSlot Checking gcd('foo') Executing query select gcd('foo') on backward Skipping gcd Checking toDayOfMonth('foo') Executing query select toDayOfMonth('foo') on backward Skipping toDayOfMonth Checking JSON_EXISTS('foo') Executing query select JSON_EXISTS('foo') on backward Skipping JSON_EXISTS Checking isIPAddressInRange('foo') Executing query select isIPAddressInRange('foo') on backward Skipping isIPAddressInRange Checking toSecond('foo') Executing query select toSecond('foo') on backward Skipping toSecond Checking hiveHash('foo') Executing query select hiveHash('foo') on backward Executing query select hiveHash('foo') on upstream OK hiveHash Checking randomStringUTF8('foo') Executing query select randomStringUTF8('foo') on backward Skipping randomStringUTF8 Checking arrayLastOrNull('foo') Executing query select arrayLastOrNull('foo') on backward Executing query select 20 on node Skipping arrayLastOrNull Checking toStartOfMonth('foo') Executing query select toStartOfMonth('foo') on backward Skipping toStartOfMonth Checking toStartOfDay('foo') Executing query select toStartOfDay('foo') on backward Skipping toStartOfDay Checking caseWithoutExpr('foo') Executing query select caseWithoutExpr('foo') on backward Skipping caseWithoutExpr Checking floor('foo') Executing query select floor('foo') on backward Skipping floor Checking bitTestAny('foo') Executing query select bitTestAny('foo') on backward Skipping bitTestAny Checking least('foo') Executing query select least('foo') on backward Executing query select least('foo') on upstream OK least Checking map('foo') Executing query select map('foo') on backward Executing query select 20 on node Skipping map Checking wordShingleMinHashCaseInsensitiveUTF8('foo') Executing query select wordShingleMinHashCaseInsensitiveUTF8('foo') on backward Executing query select wordShingleMinHashCaseInsensitiveUTF8('foo') on upstream OK wordShingleMinHashCaseInsensitiveUTF8 Checking s2CapUnion('foo') Executing query select s2CapUnion('foo') on backward Skipping s2CapUnion Checking serverUUID('foo') Executing query select serverUUID('foo') on backward Skipping serverUUID Checking toUInt256OrDefault('foo') Executing query select toUInt256OrDefault('foo') on backward Executing query select toUInt256OrDefault('foo') on upstream OK toUInt256OrDefault Checking polygonConvexHullCartesian('foo') Executing query select polygonConvexHullCartesian('foo') on backward Skipping polygonConvexHullCartesian Checking murmurHash3_32('foo') Executing query select murmurHash3_32('foo') on backward Executing query select murmurHash3_32('foo') on upstream Executing query select 20 on node OK murmurHash3_32 Checking countSubstringsCaseInsensitiveUTF8('foo') Executing query select countSubstringsCaseInsensitiveUTF8('foo') on backward Skipping countSubstringsCaseInsensitiveUTF8 Checking h3ExactEdgeLengthM('foo') Executing query select h3ExactEdgeLengthM('foo') on backward Skipping h3ExactEdgeLengthM Checking arrayFlatten('foo') Executing query select arrayFlatten('foo') on backward Skipping arrayFlatten Checking dateName('foo') Executing query select dateName('foo') on backward Skipping dateName Checking substringUTF8('foo') Executing query select substringUTF8('foo') on backward Skipping substringUTF8 Checking h3UnidirectionalEdgeIsValid('foo') Executing query select h3UnidirectionalEdgeIsValid('foo') on backward Skipping h3UnidirectionalEdgeIsValid Checking toDate('foo') Executing query select toDate('foo') on backward Skipping toDate Checking s2RectContains('foo') Executing query select s2RectContains('foo') on backward Skipping s2RectContains Checking addQuarters('foo') Executing query select addQuarters('foo') on backward Skipping addQuarters Checking cosineDistance('foo') Executing query select cosineDistance('foo') on backward Executing query select 20 on node Skipping cosineDistance Checking right('foo') Executing query select right('foo') on backward Skipping right Checking readWKTPolygon('foo') Executing query select readWKTPolygon('foo') on backward Skipping readWKTPolygon Checking polygonsSymDifferenceSpherical('foo') Executing query select polygonsSymDifferenceSpherical('foo') on backward Skipping polygonsSymDifferenceSpherical Checking tanh('foo') Executing query select tanh('foo') on backward Skipping tanh Checking dateDiff('foo') Executing query select dateDiff('foo') on backward Skipping dateDiff Checking bitRotateRight('foo') Executing query select bitRotateRight('foo') on backward Skipping bitRotateRight Checking arrayReverseSort('foo') Executing query select arrayReverseSort('foo') on backward Skipping arrayReverseSort Checking hostName('foo') Executing query select hostName('foo') on backward Skipping hostName Checking extractURLParameter('foo') Executing query select extractURLParameter('foo') on backward Skipping extractURLParameter Checking zookeeperSessionUptime('foo') Executing query select zookeeperSessionUptime('foo') on backward Skipping zookeeperSessionUptime Checking h3ToParent('foo') Executing query select h3ToParent('foo') on backward run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Skipping h3ToParent Checking arrayPartialReverseSort('foo') Executing query select arrayPartialReverseSort('foo') on backward Current start attempt failed. Will kill 5833 just in case. run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:True cmd: ['bash', '-c', 'kill -9 5833'] Command:[docker exec -u root roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c kill -9 5833] Skipping arrayPartialReverseSort Checking decodeURLComponent('foo') Executing query select decodeURLComponent('foo') on backward Stderr:bash: line 0: kill: (5833) - No such process Exitcode:1 Executing query select decodeURLComponent('foo') on upstream OK decodeURLComponent Checking bitAnd('foo') Executing query select bitAnd('foo') on backward Skipping bitAnd Checking dictGetUInt64OrDefault('foo') Executing query select dictGetUInt64OrDefault('foo') on backward Skipping dictGetUInt64OrDefault Checking firstSignificantSubdomainRFC('foo') Executing query select firstSignificantSubdomainRFC('foo') on backward Executing query select firstSignificantSubdomainRFC('foo') on upstream OK firstSignificantSubdomainRFC Checking queryID('foo') Executing query select queryID('foo') on backward Skipping queryID Checking path('foo') Executing query select path('foo') on backward Executing query select path('foo') on upstream OK path Checking defaultValueOfTypeName('foo') Executing query select defaultValueOfTypeName('foo') on backward run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Skipping defaultValueOfTypeName Checking tuplePlus('foo') Executing query select tuplePlus('foo') on backward No clickhouse process running. Start new one. http://localhost:None "POST /v1.46/containers/roottestbackwardcompatibilityconvertordinary-gw5-node-1/exec HTTP/1.1" 201 74 http://localhost:None "POST /v1.46/exec/8ea7fdaa0c36d3f2803f2cc88e0b96248779da3ef20960049e0d3bbd1e1501ec/start HTTP/1.1" 200 0 Skipping tuplePlus Checking wordShingleMinHashArgCaseInsensitive('foo') Executing query select wordShingleMinHashArgCaseInsensitive('foo') on backward http://localhost:None "GET /v1.46/exec/8ea7fdaa0c36d3f2803f2cc88e0b96248779da3ef20960049e0d3bbd1e1501ec/json HTTP/1.1" 200 587 Executing query select wordShingleMinHashArgCaseInsensitive('foo') on upstream OK wordShingleMinHashArgCaseInsensitive Checking subtractWeeks('foo') Executing query select subtractWeeks('foo') on backward Skipping subtractWeeks Checking arrayFirst('foo') Executing query select arrayFirst('foo') on backward Skipping arrayFirst Checking accurateCastOrDefault('foo') Executing query select accurateCastOrDefault('foo') on backward Skipping accurateCastOrDefault Checking stem('foo') Executing query select stem('foo') on backward Skipping stem Checking arrayReverseSplit('foo') Executing query select arrayReverseSplit('foo') on backward Skipping arrayReverseSplit Checking transactionOldestSnapshot('foo') Executing query select transactionOldestSnapshot('foo') on backward Skipping transactionOldestSnapshot Checking dumpColumnStructure('foo') Executing query select dumpColumnStructure('foo') on backward Executing query select dumpColumnStructure('foo') on upstream OK dumpColumnStructure Checking multiSearchFirstIndex('foo') Executing query select multiSearchFirstIndex('foo') on backward Skipping multiSearchFirstIndex Checking tcpPort('foo') Executing query select tcpPort('foo') on backward Skipping tcpPort Checking subtractHours('foo') Executing query select subtractHours('foo') on backward Skipping subtractHours Checking parseDateTimeBestEffortUS('foo') Executing query select parseDateTimeBestEffortUS('foo') on backward Skipping parseDateTimeBestEffortUS Checking mapPopulateSeries('foo') Executing query select mapPopulateSeries('foo') on backward Skipping mapPopulateSeries Checking parseDateTime64BestEffortUS('foo') Executing query select parseDateTime64BestEffortUS('foo') on backward Skipping parseDateTime64BestEffortUS Checking synonyms('foo') Executing query select synonyms('foo') on backward Skipping synonyms Checking isFinite('foo') Executing query select isFinite('foo') on backward Skipping isFinite Checking toDayOfWeek('foo') Executing query select toDayOfWeek('foo') on backward Skipping toDayOfWeek Checking reinterpretAsString('foo') Executing query select reinterpretAsString('foo') on backward Executing query select reinterpretAsString('foo') on upstream run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] OK reinterpretAsString Checking trimLeft('foo') Executing query select trimLeft('foo') on backward Stdout:6590 Clickhouse process running. run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select trimLeft('foo') on upstream Stdout:6590 Executing query select 20 on node OK trimLeft Checking subtractInterval('foo') Executing query select subtractInterval('foo') on backward Skipping subtractInterval Checking toValidUTF8('foo') Executing query select toValidUTF8('foo') on backward Executing query select toValidUTF8('foo') on upstream OK toValidUTF8 Checking abs('foo') Executing query select abs('foo') on backward Skipping abs Checking arrayWithConstant('foo') Executing query select arrayWithConstant('foo') on backward Skipping arrayWithConstant Checking toStartOfNanosecond('foo') Executing query select toStartOfNanosecond('foo') on backward Skipping toStartOfNanosecond Checking toIntervalNanosecond('foo') Executing query select toIntervalNanosecond('foo') on backward Connection dropped: socket connection error: No route to host Connection dropped: socket connection error: No route to host Skipping toIntervalNanosecond Checking SHA256('foo') Executing query select SHA256('foo') on backward Executing query select SHA256('foo') on upstream OK SHA256 Checking LpDistance('foo') Executing query select LpDistance('foo') on backward Executing query select 20 on node Skipping LpDistance Checking arrayMax('foo') Executing query select arrayMax('foo') on backward Skipping arrayMax Checking toUUIDOrDefault('foo') Executing query select toUUIDOrDefault('foo') on backward Executing query select toUUIDOrDefault('foo') on upstream OK toUUIDOrDefault Checking LinfNormalize('foo') Executing query select LinfNormalize('foo') on backward Skipping LinfNormalize Checking extractURLParameterNames('foo') Executing query select extractURLParameterNames('foo') on backward Executing query select extractURLParameterNames('foo') on upstream OK extractURLParameterNames Checking dictGetInt8('foo') Executing query select dictGetInt8('foo') on backward Skipping dictGetInt8 Checking h3PointDistKm('foo') Executing query select h3PointDistKm('foo') on backward Skipping h3PointDistKm Checking multiSearchAllPositionsUTF8('foo') Executing query select multiSearchAllPositionsUTF8('foo') on backward Executing query select 20 on node Skipping multiSearchAllPositionsUTF8 Checking toYearWeek('foo') Executing query select toYearWeek('foo') on backward Skipping toYearWeek Checking toFixedString('foo') Executing query select toFixedString('foo') on backward Skipping toFixedString Checking extractAllGroupsVertical('foo') Executing query select extractAllGroupsVertical('foo') on backward Skipping extractAllGroupsVertical Checking MACNumToString('foo') Executing query select MACNumToString('foo') on backward Skipping MACNumToString Checking revision('foo') Executing query select revision('foo') on backward Skipping revision Checking hasTokenCaseInsensitive('foo') Executing query select hasTokenCaseInsensitive('foo') on backward Skipping hasTokenCaseInsensitive Checking toStartOfYear('foo') Executing query select toStartOfYear('foo') on backward Skipping toStartOfYear Checking toDecimal256('foo') Executing query select toDecimal256('foo') on backward Skipping toDecimal256 Checking multiSearchFirstPositionCaseInsensitiveUTF8('foo') Executing query select multiSearchFirstPositionCaseInsensitiveUTF8('foo') on backward Executing query select 20 on node Skipping multiSearchFirstPositionCaseInsensitiveUTF8 Checking cutFragment('foo') Executing query select cutFragment('foo') on backward Executing query select cutFragment('foo') on upstream OK cutFragment Checking arrayCompact('foo') Executing query select arrayCompact('foo') on backward Skipping arrayCompact Checking toIntervalWeek('foo') Executing query select toIntervalWeek('foo') on backward Skipping toIntervalWeek Checking toRelativeSecondNum('foo') Executing query select toRelativeSecondNum('foo') on backward Executing query SELECT uuid FROM system.tables WHERE name='tbl' on node Skipping toRelativeSecondNum Checking replaceAll('foo') Executing query select replaceAll('foo') on backward Executing query ALTER TABLE tbl FREEZE WITH NAME 'test' on node run container_id:roottestcancelfreeze-gw1-node-1 detach:False nothrow:True cmd: ['bash', '-c', 'ls /var/lib/clickhouse/shadow/'] Command:[docker exec roottestcancelfreeze-gw1-node-1 bash -c ls /var/lib/clickhouse/shadow/] Skipping replaceAll Checking bitShiftLeft('foo') Executing query select bitShiftLeft('foo') on backward Stderr:ls: cannot access '/var/lib/clickhouse/shadow/': No such file or directory Exitcode:2 run container_id:roottestcancelfreeze-gw1-node-1 detach:False nothrow:True cmd: ['bash', '-c', 'ls /var/lib/clickhouse/shadow/'] Command:[docker exec roottestcancelfreeze-gw1-node-1 bash -c ls /var/lib/clickhouse/shadow/] Skipping bitShiftLeft Checking netloc('foo') Executing query select netloc('foo') on backward Stdout:increment.txt Stdout:test Executing query SELECT count() FROM system.processes WHERE query_kind == 'Alter' AND query LIKE '%FREEZE%' on node Executing query select 20 on node Executing query select netloc('foo') on upstream Executing query KILL QUERY WHERE query_kind == 'Alter' AND query LIKE '%FREEZE%' SYNC on node OK netloc Checking dictGetIPv4OrDefault('foo') Executing query select dictGetIPv4OrDefault('foo') on backward Skipping dictGetIPv4OrDefault Checking arrayEnumerate('foo') Executing query select arrayEnumerate('foo') on backward Skipping arrayEnumerate Checking exp('foo') Executing query select exp('foo') on backward run container_id:roottestcancelfreeze-gw1-node-1 detach:False nothrow:False cmd: ['bash', '-c', 'ls /var/lib/clickhouse/shadow/test/store/e83/e83f02d2-31d1-4d59-99db-37b63db502e1'] Command:[docker exec roottestcancelfreeze-gw1-node-1 bash -c ls /var/lib/clickhouse/shadow/test/store/e83/e83f02d2-31d1-4d59-99db-37b63db502e1] Skipping exp Checking tupleNegate('foo') Executing query select tupleNegate('foo') on backward Stdout:0_1_1_0 Stdout:10000_10001_10001_0 Stdout:1000_1001_1001_0 Stdout:10001_10002_10002_0 Stdout:10002_10003_10003_0 Stdout:10003_10004_10004_0 Stdout:10004_10005_10005_0 Stdout:10005_10006_10006_0 Stdout:10006_10007_10007_0 Stdout:10007_10008_10008_0 Stdout:10008_10009_10009_0 Stdout:10009_10010_10010_0 Stdout:10010_10011_10011_0 Stdout:100_101_101_0 Stdout:1001_1002_1002_0 Stdout:10011_10012_10012_0 Stdout:10012_10013_10013_0 Stdout:10013_10014_10014_0 Stdout:10014_10015_10015_0 Stdout:10015_10016_10016_0 Stdout:10016_10017_10017_0 Stdout:10017_10018_10018_0 Stdout:10018_10019_10019_0 Stdout:10019_10020_10020_0 Stdout:10020_10021_10021_0 Stdout:1002_1003_1003_0 Stdout:10021_10022_10022_0 Stdout:10022_10023_10023_0 Stdout:10023_10024_10024_0 Stdout:10024_10025_10025_0 Stdout:10025_10026_10026_0 Stdout:10026_10027_10027_0 Stdout:10027_10028_10028_0 Stdout:10028_10029_10029_0 Stdout:10029_10030_10030_0 Stdout:10030_10031_10031_0 Stdout:1003_1004_1004_0 Stdout:10031_10032_10032_0 Stdout:10032_10033_10033_0 Stdout:10033_10034_10034_0 Stdout:10034_10035_10035_0 Stdout:10035_10036_10036_0 Stdout:10036_10037_10037_0 Stdout:10037_10038_10038_0 Stdout:10038_10039_10039_0 Stdout:10039_10040_10040_0 Stdout:10040_10041_10041_0 Stdout:1004_1005_1005_0 Stdout:10041_10042_10042_0 Stdout:10042_10043_10043_0 Stdout:10043_10044_10044_0 Stdout:10044_10045_10045_0 Stdout:10045_10046_10046_0 Stdout:10046_10047_10047_0 Stdout:10047_10048_10048_0 Stdout:10048_10049_10049_0 Stdout:10049_10050_10050_0 Stdout:10050_10051_10051_0 Stdout:1005_1006_1006_0 Stdout:10051_10052_10052_0 Stdout:10052_10053_10053_0 Stdout:10053_10054_10054_0 Stdout:10054_10055_10055_0 Stdout:10055_10056_10056_0 Stdout:10056_10057_10057_0 Stdout:10057_10058_10058_0 Stdout:10058_10059_10059_0 Stdout:10059_10060_10060_0 Stdout:10060_10061_10061_0 Stdout:1006_1007_1007_0 Stdout:10061_10062_10062_0 Stdout:10062_10063_10063_0 Stdout:10063_10064_10064_0 Stdout:10064_10065_10065_0 Stdout:10065_10066_10066_0 Stdout:10066_10067_10067_0 Stdout:10067_10068_10068_0 Stdout:10068_10069_10069_0 Stdout:10069_10070_10070_0 Stdout:10070_10071_10071_0 Stdout:1007_1008_1008_0 Stdout:10071_10072_10072_0 Stdout:10072_10073_10073_0 Stdout:10073_10074_10074_0 Stdout:10074_10075_10075_0 Stdout:10075_10076_10076_0 Stdout:10076_10077_10077_0 Stdout:10077_10078_10078_0 Stdout:10078_10079_10079_0 Stdout:10079_10080_10080_0 Stdout:10080_10081_10081_0 Stdout:1008_1009_1009_0 Stdout:10081_10082_10082_0 Stdout:10082_10083_10083_0 Stdout:10083_10084_10084_0 Stdout:10084_10085_10085_0 Stdout:10085_10086_10086_0 Stdout:10086_10087_10087_0 Stdout:10087_10088_10088_0 Stdout:10088_10089_10089_0 Stdout:10089_10090_10090_0 Stdout:10090_10091_10091_0 Stdout:1009_1010_1010_0 Stdout:10091_10092_10092_0 Stdout:10092_10093_10093_0 Stdout:10093_10094_10094_0 Stdout:10094_10095_10095_0 Stdout:10095_10096_10096_0 Stdout:10096_10097_10097_0 Stdout:10097_10098_10098_0 Stdout:10098_10099_10099_0 Stdout:10099_10100_10100_0 Stdout:10100_10101_10101_0 Stdout:1010_1011_1011_0 Stdout:10101_10102_10102_0 Stdout:10102_10103_10103_0 Stdout:10103_10104_10104_0 Stdout:10104_10105_10105_0 Stdout:10105_10106_10106_0 Stdout:10106_10107_10107_0 Stdout:10107_10108_10108_0 Stdout:10108_10109_10109_0 Stdout:10109_10110_10110_0 Stdout:10110_10111_10111_0 Stdout:101_102_102_0 Stdout:1011_1012_1012_0 Stdout:10_11_11_0 Stdout:10111_10112_10112_0 Stdout:10112_10113_10113_0 Stdout:10113_10114_10114_0 Stdout:10114_10115_10115_0 Stdout:10115_10116_10116_0 Stdout:10116_10117_10117_0 Stdout:10117_10118_10118_0 Stdout:10118_10119_10119_0 Stdout:10119_10120_10120_0 Stdout:10120_10121_10121_0 Stdout:1012_1013_1013_0 Stdout:10121_10122_10122_0 Stdout:10122_10123_10123_0 Stdout:10123_10124_10124_0 Stdout:10124_10125_10125_0 Stdout:10125_10126_10126_0 Stdout:10126_10127_10127_0 Stdout:10127_10128_10128_0 Stdout:10128_10129_10129_0 Stdout:10129_10130_10130_0 Stdout:10130_10131_10131_0 Stdout:1013_1014_1014_0 Stdout:10131_10132_10132_0 Stdout:10132_10133_10133_0 Stdout:10133_10134_10134_0 Stdout:10134_10135_10135_0 Stdout:10135_10136_10136_0 Stdout:10136_10137_10137_0 Stdout:10137_10138_10138_0 Stdout:10138_10139_10139_0 Stdout:10139_10140_10140_0 Stdout:10140_10141_10141_0 Stdout:1014_1015_1015_0 Stdout:10141_10142_10142_0 Stdout:10142_10143_10143_0 Stdout:10143_10144_10144_0 Stdout:10144_10145_10145_0 Stdout:10145_10146_10146_0 Stdout:10146_10147_10147_0 Stdout:10147_10148_10148_0 Stdout:10148_10149_10149_0 Stdout:10149_10150_10150_0 Stdout:10150_10151_10151_0 Stdout:1015_1016_1016_0 Stdout:10151_10152_10152_0 Stdout:10152_10153_10153_0 Stdout:10153_10154_10154_0 Stdout:10154_10155_10155_0 Stdout:10155_10156_10156_0 Stdout:10156_10157_10157_0 Stdout:10157_10158_10158_0 Stdout:10158_10159_10159_0 Stdout:10159_10160_10160_0 Stdout:10160_10161_10161_0 Stdout:1016_1017_1017_0 Stdout:10161_10162_10162_0 Stdout:10162_10163_10163_0 Stdout:10163_10164_10164_0 Stdout:10164_10165_10165_0 Stdout:10165_10166_10166_0 Stdout:10166_10167_10167_0 Stdout:10167_10168_10168_0 Stdout:10168_10169_10169_0 Stdout:10169_10170_10170_0 Stdout:10170_10171_10171_0 Stdout:1017_1018_1018_0 Stdout:10171_10172_10172_0 Stdout:10172_10173_10173_0 Stdout:10173_10174_10174_0 Stdout:10174_10175_10175_0 Stdout:10175_10176_10176_0 Stdout:10176_10177_10177_0 Stdout:10177_10178_10178_0 Stdout:10178_10179_10179_0 Stdout:10179_10180_10180_0 Stdout:10180_10181_10181_0 Stdout:1018_1019_1019_0 Stdout:10181_10182_10182_0 Stdout:10182_10183_10183_0 Stdout:10183_10184_10184_0 Stdout:10184_10185_10185_0 Stdout:10185_10186_10186_0 Stdout:10186_10187_10187_0 Stdout:10187_10188_10188_0 Stdout:10188_10189_10189_0 Stdout:10189_10190_10190_0 Stdout:10190_10191_10191_0 Stdout:1019_1020_1020_0 Stdout:10191_10192_10192_0 Stdout:10192_10193_10193_0 Stdout:10193_10194_10194_0 Stdout:10194_10195_10195_0 Stdout:10195_10196_10196_0 Stdout:10196_10197_10197_0 Stdout:10197_10198_10198_0 Stdout:10198_10199_10199_0 Stdout:10199_10200_10200_0 Stdout:10200_10201_10201_0 Stdout:1020_1021_1021_0 Stdout:10201_10202_10202_0 Stdout:10202_10203_10203_0 Stdout:10203_10204_10204_0 Stdout:10204_10205_10205_0 Stdout:10205_10206_10206_0 Stdout:10206_10207_10207_0 Stdout:10207_10208_10208_0 Stdout:10208_10209_10209_0 Stdout:10209_10210_10210_0 Stdout:10210_10211_10211_0 Stdout:102_103_103_0 Stdout:1021_1022_1022_0 Stdout:10211_10212_10212_0 Stdout:10212_10213_10213_0 Stdout:10213_10214_10214_0 Stdout:10214_10215_10215_0 Stdout:10215_10216_10216_0 Stdout:10216_10217_10217_0 Stdout:10217_10218_10218_0 Stdout:10218_10219_10219_0 Stdout:10219_10220_10220_0 Stdout:10220_10221_10221_0 Stdout:1022_1023_1023_0 Stdout:10221_10222_10222_0 Stdout:10222_10223_10223_0 Stdout:10223_10224_10224_0 Stdout:10224_10225_10225_0 Stdout:10225_10226_10226_0 Stdout:10226_10227_10227_0 Stdout:10227_10228_10228_0 Stdout:10228_10229_10229_0 Stdout:10229_10230_10230_0 Stdout:10230_10231_10231_0 Stdout:1023_1024_1024_0 Stdout:10231_10232_10232_0 Stdout:10232_10233_10233_0 Stdout:10233_10234_10234_0 Stdout:10234_10235_10235_0 Stdout:10235_10236_10236_0 Stdout:10236_10237_10237_0 Stdout:10237_10238_10238_0 Stdout:10238_10239_10239_0 Stdout:10239_10240_10240_0 Stdout:10240_10241_10241_0 Stdout:1024_1025_1025_0 Stdout:10241_10242_10242_0 Stdout:10242_10243_10243_0 Stdout:10243_10244_10244_0 Stdout:10244_10245_10245_0 Stdout:10245_10246_10246_0 Stdout:10246_10247_10247_0 Stdout:10247_10248_10248_0 Stdout:10248_10249_10249_0 Stdout:10249_10250_10250_0 Stdout:10250_10251_10251_0 Stdout:1025_1026_1026_0 Stdout:10251_10252_10252_0 Stdout:10252_10253_10253_0 Stdout:10253_10254_10254_0 Stdout:10254_10255_10255_0 Stdout:10255_10256_10256_0 Stdout:10256_10257_10257_0 Stdout:10257_10258_10258_0 Stdout:10258_10259_10259_0 Stdout:10259_10260_10260_0 Stdout:10260_10261_10261_0 Stdout:1026_1027_1027_0 Stdout:10261_10262_10262_0 Stdout:10262_10263_10263_0 Stdout:10263_10264_10264_0 Stdout:10264_10265_10265_0 Stdout:10265_10266_10266_0 Stdout:10266_10267_10267_0 Stdout:10267_10268_10268_0 Stdout:10268_10269_10269_0 Stdout:10269_10270_10270_0 Stdout:10270_10271_10271_0 Stdout:1027_1028_1028_0 Stdout:10271_10272_10272_0 Stdout:10272_10273_10273_0 Stdout:10273_10274_10274_0 Stdout:10274_10275_10275_0 Stdout:10275_10276_10276_0 Stdout:10276_10277_10277_0 Stdout:10277_10278_10278_0 Stdout:10278_10279_10279_0 Stdout:10279_10280_10280_0 Stdout:10280_10281_10281_0 Stdout:1028_1029_1029_0 Stdout:10281_10282_10282_0 Stdout:10282_10283_10283_0 Stdout:10283_10284_10284_0 Stdout:10284_10285_10285_0 Stdout:10285_10286_10286_0 Stdout:10286_10287_10287_0 Stdout:10287_10288_10288_0 Stdout:10288_10289_10289_0 Stdout:10289_10290_10290_0 Stdout:10290_10291_10291_0 Stdout:1029_1030_1030_0 Stdout:10291_10292_10292_0 Stdout:10292_10293_10293_0 Stdout:10293_10294_10294_0 Stdout:10294_10295_10295_0 Stdout:10295_10296_10296_0 Stdout:10296_10297_10297_0 Stdout:10297_10298_10298_0 Stdout:10298_10299_10299_0 Stdout:10299_10300_10300_0 Stdout:10300_10301_10301_0 Stdout:1030_1031_1031_0 Stdout:10301_10302_10302_0 Stdout:10302_10303_10303_0 Stdout:10303_10304_10304_0 Stdout:10304_10305_10305_0 Stdout:10305_10306_10306_0 Stdout:10306_10307_10307_0 Stdout:10307_10308_10308_0 Stdout:10308_10309_10309_0 Stdout:10309_10310_10310_0 Stdout:10310_10311_10311_0 Stdout:103_104_104_0 Stdout:1031_1032_1032_0 Stdout:10311_10312_10312_0 Stdout:10312_10313_10313_0 Stdout:10313_10314_10314_0 Stdout:10314_10315_10315_0 Stdout:10315_10316_10316_0 Stdout:10316_10317_10317_0 Stdout:10317_10318_10318_0 Stdout:10318_10319_10319_0 Stdout:10319_10320_10320_0 Stdout:10320_10321_10321_0 Stdout:1032_1033_1033_0 Stdout:10321_10322_10322_0 Stdout:10322_10323_10323_0 Stdout:10323_10324_10324_0 Stdout:10324_10325_10325_0 Stdout:10325_10326_10326_0 Stdout:10326_10327_10327_0 Stdout:10327_10328_10328_0 Stdout:10328_10329_10329_0 Stdout:10329_10330_10330_0 Stdout:10330_10331_10331_0 Stdout:1033_1034_1034_0 Stdout:10331_10332_10332_0 Stdout:10332_10333_10333_0 Stdout:10333_10334_10334_0 Stdout:10334_10335_10335_0 Stdout:10335_10336_10336_0 Stdout:10336_10337_10337_0 Stdout:10337_10338_10338_0 Stdout:10338_10339_10339_0 Stdout:10339_10340_10340_0 Stdout:10340_10341_10341_0 Stdout:1034_1035_1035_0 Stdout:10341_10342_10342_0 Stdout:10342_10343_10343_0 Stdout:10343_10344_10344_0 Stdout:10344_10345_10345_0 Stdout:10345_10346_10346_0 Stdout:10346_10347_10347_0 Stdout:10347_10348_10348_0 Stdout:10348_10349_10349_0 Stdout:10349_10350_10350_0 Stdout:10350_10351_10351_0 Stdout:1035_1036_1036_0 Stdout:10351_10352_10352_0 Stdout:10352_10353_10353_0 Stdout:10353_10354_10354_0 Stdout:10354_10355_10355_0 Stdout:10355_10356_10356_0 Stdout:10356_10357_10357_0 Stdout:10357_10358_10358_0 Stdout:10358_10359_10359_0 Stdout:10359_10360_10360_0 Stdout:10360_10361_10361_0 Stdout:1036_1037_1037_0 Stdout:10361_10362_10362_0 Stdout:10362_10363_10363_0 Stdout:10363_10364_10364_0 Stdout:10364_10365_10365_0 Stdout:10365_10366_10366_0 Stdout:10366_10367_10367_0 Stdout:10367_10368_10368_0 Stdout:10368_10369_10369_0 Stdout:10369_10370_10370_0 Stdout:10370_10371_10371_0 Stdout:1037_1038_1038_0 Stdout:10371_10372_10372_0 Stdout:10372_10373_10373_0 Stdout:10373_10374_10374_0 Stdout:10374_10375_10375_0 Stdout:10375_10376_10376_0 Stdout:10376_10377_10377_0 Stdout:10377_10378_10378_0 Stdout:10378_10379_10379_0 Stdout:10379_10380_10380_0 Stdout:10380_10381_10381_0 Stdout:1038_1039_1039_0 Stdout:10381_10382_10382_0 Stdout:10382_10383_10383_0 Stdout:10383_10384_10384_0 Stdout:10384_10385_10385_0 Stdout:10385_10386_10386_0 Stdout:10386_10387_10387_0 Stdout:10387_10388_10388_0 Stdout:10388_10389_10389_0 Stdout:10389_10390_10390_0 Stdout:10390_10391_10391_0 Stdout:1039_1040_1040_0 Stdout:10391_10392_10392_0 Stdout:10392_10393_10393_0 Stdout:10393_10394_10394_0 Stdout:10394_10395_10395_0 Stdout:10395_10396_10396_0 Stdout:10396_10397_10397_0 Stdout:10397_10398_10398_0 Stdout:10398_10399_10399_0 Stdout:10399_10400_10400_0 Stdout:10400_10401_10401_0 Stdout:1040_1041_1041_0 Stdout:10401_10402_10402_0 Stdout:10402_10403_10403_0 Stdout:10403_10404_10404_0 Stdout:10404_10405_10405_0 Stdout:10405_10406_10406_0 Stdout:10406_10407_10407_0 Stdout:10407_10408_10408_0 Stdout:10408_10409_10409_0 Stdout:10409_10410_10410_0 Stdout:10410_10411_10411_0 Stdout:104_105_105_0 Stdout:1041_1042_1042_0 Stdout:10411_10412_10412_0 Stdout:10412_10413_10413_0 Stdout:10413_10414_10414_0 Stdout:10414_10415_10415_0 Stdout:10415_10416_10416_0 Stdout:10416_10417_10417_0 Stdout:10417_10418_10418_0 Stdout:10418_10419_10419_0 Stdout:10419_10420_10420_0 Stdout:10420_10421_10421_0 Stdout:1042_1043_1043_0 Stdout:10421_10422_10422_0 Stdout:10422_10423_10423_0 Stdout:10423_10424_10424_0 Stdout:10424_10425_10425_0 Stdout:10425_10426_10426_0 Stdout:10426_10427_10427_0 Stdout:10427_10428_10428_0 Stdout:10428_10429_10429_0 Stdout:10429_10430_10430_0 Stdout:10430_10431_10431_0 Stdout:1043_1044_1044_0 Stdout:10431_10432_10432_0 Stdout:10432_10433_10433_0 Stdout:10433_10434_10434_0 Stdout:10434_10435_10435_0 Stdout:10435_10436_10436_0 Stdout:10436_10437_10437_0 Stdout:10437_10438_10438_0 Stdout:10438_10439_10439_0 Stdout:10439_10440_10440_0 Stdout:10440_10441_10441_0 Stdout:1044_1045_1045_0 Stdout:10441_10442_10442_0 Stdout:10442_10443_10443_0 Stdout:10443_10444_10444_0 Stdout:10444_10445_10445_0 Stdout:10445_10446_10446_0 Stdout:10446_10447_10447_0 Stdout:10447_10448_10448_0 Stdout:10448_10449_10449_0 Stdout:10449_10450_10450_0 Stdout:10450_10451_10451_0 Stdout:1045_1046_1046_0 Stdout:10451_10452_10452_0 Stdout:10452_10453_10453_0 Stdout:10453_10454_10454_0 Stdout:10454_10455_10455_0 Stdout:10455_10456_10456_0 Stdout:10456_10457_10457_0 Stdout:10457_10458_10458_0 Stdout:10458_10459_10459_0 Stdout:10459_10460_10460_0 Stdout:10460_10461_10461_0 Stdout:1046_1047_1047_0 Stdout:10461_10462_10462_0 Stdout:10462_10463_10463_0 Stdout:10463_10464_10464_0 Stdout:10464_10465_10465_0 Stdout:10465_10466_10466_0 Stdout:10466_10467_10467_0 Stdout:10467_10468_10468_0 Stdout:10468_10469_10469_0 Stdout:10469_10470_10470_0 Stdout:10470_10471_10471_0 Stdout:1047_1048_1048_0 Stdout:10471_10472_10472_0 Stdout:10472_10473_10473_0 Stdout:10473_10474_10474_0 Stdout:10474_10475_10475_0 Stdout:10475_10476_10476_0 Stdout:10476_10477_10477_0 Stdout:10477_10478_10478_0 Stdout:10478_10479_10479_0 Stdout:10479_10480_10480_0 Stdout:10480_10481_10481_0 Stdout:1048_1049_1049_0 Stdout:10481_10482_10482_0 Stdout:10482_10483_10483_0 Stdout:10483_10484_10484_0 Stdout:10484_10485_10485_0 Stdout:10485_10486_10486_0 Stdout:10486_10487_10487_0 Stdout:10487_10488_10488_0 Stdout:10488_10489_10489_0 Stdout:10489_10490_10490_0 Stdout:10490_10491_10491_0 Stdout:1049_1050_1050_0 Stdout:10491_10492_10492_0 Stdout:10492_10493_10493_0 Stdout:10493_10494_10494_0 Stdout:10494_10495_10495_0 Stdout:10495_10496_10496_0 Stdout:10496_10497_10497_0 Stdout:10497_10498_10498_0 Stdout:10498_10499_10499_0 Stdout:10499_10500_10500_0 Stdout:10500_10501_10501_0 Stdout:1050_1051_1051_0 Stdout:10501_10502_10502_0 Stdout:10502_10503_10503_0 Stdout:10503_10504_10504_0 Stdout:10504_10505_10505_0 Stdout:10505_10506_10506_0 Stdout:10506_10507_10507_0 Stdout:10507_10508_10508_0 Stdout:10508_10509_10509_0 Stdout:10509_10510_10510_0 Stdout:10510_10511_10511_0 Stdout:105_106_106_0 Stdout:1051_1052_1052_0 Stdout:10511_10512_10512_0 Stdout:10512_10513_10513_0 Stdout:10513_10514_10514_0 Stdout:10514_10515_10515_0 Stdout:10515_10516_10516_0 Stdout:10516_10517_10517_0 Stdout:10517_10518_10518_0 Stdout:10518_10519_10519_0 Stdout:10519_10520_10520_0 Stdout:10520_10521_10521_0 Stdout:1052_1053_1053_0 Stdout:10521_10522_10522_0 Stdout:10522_10523_10523_0 Stdout:10523_10524_10524_0 Stdout:10524_10525_10525_0 Stdout:10525_10526_10526_0 Stdout:10526_10527_10527_0 Stdout:10527_10528_10528_0 Stdout:10528_10529_10529_0 Stdout:10529_10530_10530_0 Stdout:10530_10531_10531_0 Stdout:1053_1054_1054_0 Stdout:10531_10532_10532_0 Stdout:10532_10533_10533_0 Stdout:10533_10534_10534_0 Stdout:10534_10535_10535_0 Stdout:10535_10536_10536_0 Stdout:10536_10537_10537_0 Stdout:10537_10538_10538_0 Stdout:10538_10539_10539_0 Stdout:10539_10540_10540_0 Stdout:10540_10541_10541_0 Stdout:1054_1055_1055_0 Stdout:10541_10542_10542_0 Stdout:10542_10543_10543_0 Stdout:10543_10544_10544_0 Stdout:10544_10545_10545_0 Stdout:10545_10546_10546_0 Stdout:10546_10547_10547_0 Stdout:10547_10548_10548_0 Stdout:10548_10549_10549_0 Stdout:10549_10550_10550_0 Stdout:10550_10551_10551_0 Stdout:1055_1056_1056_0 Stdout:10551_10552_10552_0 Stdout:10552_10553_10553_0 Stdout:10553_10554_10554_0 Stdout:10554_10555_10555_0 Stdout:10555_10556_10556_0 Stdout:10556_10557_10557_0 Stdout:10557_10558_10558_0 Stdout:10558_10559_10559_0 Stdout:10559_10560_10560_0 Stdout:10560_10561_10561_0 Stdout:1056_1057_1057_0 Stdout:10561_10562_10562_0 Stdout:10562_10563_10563_0 Stdout:10563_10564_10564_0 Stdout:10564_10565_10565_0 Stdout:10565_10566_10566_0 Stdout:10566_10567_10567_0 Stdout:10567_10568_10568_0 Stdout:10568_10569_10569_0 Stdout:10569_10570_10570_0 Stdout:10570_10571_10571_0 Stdout:1057_1058_1058_0 Stdout:10571_10572_10572_0 Stdout:10572_10573_10573_0 Stdout:10573_10574_10574_0 Stdout:10574_10575_10575_0 Stdout:10575_10576_10576_0 Stdout:10576_10577_10577_0 Stdout:10577_10578_10578_0 Stdout:10578_10579_10579_0 Stdout:10579_10580_10580_0 Stdout:10580_10581_10581_0 Stdout:1058_1059_1059_0 Stdout:10581_10582_10582_0 Stdout:10582_10583_10583_0 Stdout:10583_10584_10584_0 Stdout:10584_10585_10585_0 Stdout:10585_10586_10586_0 Stdout:10586_10587_10587_0 Stdout:10587_10588_10588_0 Stdout:10588_10589_10589_0 Stdout:10589_10590_10590_0 Stdout:10590_10591_10591_0 Stdout:1059_1060_1060_0 Stdout:10591_10592_10592_0 Stdout:10592_10593_10593_0 Stdout:10593_10594_10594_0 Stdout:10594_10595_10595_0 Stdout:10595_10596_10596_0 Stdout:10596_10597_10597_0 Stdout:10597_10598_10598_0 Stdout:10598_10599_10599_0 Stdout:10599_10600_10600_0 Stdout:10600_10601_10601_0 Stdout:1060_1061_1061_0 Stdout:10601_10602_10602_0 Stdout:10602_10603_10603_0 Stdout:10603_10604_10604_0 Stdout:10604_10605_10605_0 Stdout:10605_10606_10606_0 Stdout:10606_10607_10607_0 Stdout:10607_10608_10608_0 Stdout:10608_10609_10609_0 Stdout:10609_10610_10610_0 Stdout:10610_10611_10611_0 Stdout:106_107_107_0 Stdout:1061_1062_1062_0 Stdout:10611_10612_10612_0 Stdout:10612_10613_10613_0 Stdout:10613_10614_10614_0 Stdout:10614_10615_10615_0 Stdout:10615_10616_10616_0 Stdout:10616_10617_10617_0 Stdout:10617_10618_10618_0 Stdout:10618_10619_10619_0 Stdout:10619_10620_10620_0 Stdout:10620_10621_10621_0 Stdout:1062_1063_1063_0 Stdout:10621_10622_10622_0 Stdout:10622_10623_10623_0 Stdout:10623_10624_10624_0 Stdout:10624_10625_10625_0 Stdout:10625_10626_10626_0 Stdout:10626_10627_10627_0 Stdout:10627_10628_10628_0 Stdout:10628_10629_10629_0 Stdout:10629_10630_10630_0 Stdout:10630_10631_10631_0 Stdout:1063_1064_1064_0 Stdout:10631_10632_10632_0 Stdout:10632_10633_10633_0 Stdout:10633_10634_10634_0 Stdout:10634_10635_10635_0 Stdout:10635_10636_10636_0 Stdout:10636_10637_10637_0 Stdout:10637_10638_10638_0 Stdout:10638_10639_10639_0 Stdout:10639_10640_10640_0 Stdout:10640_10641_10641_0 Stdout:1064_1065_1065_0 Stdout:10641_10642_10642_0 Stdout:10642_10643_10643_0 Stdout:10643_10644_10644_0 Stdout:10644_10645_10645_0 Stdout:10645_10646_10646_0 Stdout:10646_10647_10647_0 Stdout:10647_10648_10648_0 Stdout:10648_10649_10649_0 Stdout:10649_10650_10650_0 Stdout:10650_10651_10651_0 Stdout:1065_1066_1066_0 Stdout:10651_10652_10652_0 Stdout:10652_10653_10653_0 Stdout:10653_10654_10654_0 Stdout:10654_10655_10655_0 Stdout:10655_10656_10656_0 Stdout:10656_10657_10657_0 Stdout:10657_10658_10658_0 Stdout:10658_10659_10659_0 Stdout:10659_10660_10660_0 Stdout:10660_10661_10661_0 Stdout:1066_1067_1067_0 Stdout:10661_10662_10662_0 Stdout:10662_10663_10663_0 Stdout:10663_10664_10664_0 Stdout:10664_10665_10665_0 Stdout:10665_10666_10666_0 Stdout:10666_10667_10667_0 Stdout:10667_10668_10668_0 Stdout:10668_10669_10669_0 Stdout:10669_10670_10670_0 Stdout:10670_10671_10671_0 Stdout:1067_1068_1068_0 Stdout:10671_10672_10672_0 Stdout:10672_10673_10673_0 Stdout:10673_10674_10674_0 Stdout:10674_10675_10675_0 Stdout:10675_10676_10676_0 Stdout:1_2_2_0 [gw1] PASSED test_cancel_freeze/test.py::test_cancel_backup Executing query DROP TABLE IF EXISTS tbl SYNC on node Skipping tupleNegate Checking toRelativeQuarterNum('foo') Executing query select toRelativeQuarterNum('foo') on backward Skipping toRelativeQuarterNum Checking widthBucket('foo') Executing query select widthBucket('foo') on backward Skipping widthBucket Checking toRelativeYearNum('foo') Executing query select toRelativeYearNum('foo') on backward Skipping toRelativeYearNum Checking h3IsResClassIII('foo') Executing query select h3IsResClassIII('foo') on backward Executing query select 20 on node Skipping h3IsResClassIII Checking cutQueryStringAndFragment('foo') Executing query select cutQueryStringAndFragment('foo') on backward Executing query select cutQueryStringAndFragment('foo') on upstream OK cutQueryStringAndFragment Checking repeat('foo') Executing query select repeat('foo') on backward Skipping repeat Checking unhex('foo') Executing query select unhex('foo') on backward Executing query select unhex('foo') on upstream OK unhex Checking IPv4ToIPv6('foo') Executing query select IPv4ToIPv6('foo') on backward Skipping IPv4ToIPv6 Checking toStartOfMicrosecond('foo') Executing query select toStartOfMicrosecond('foo') on backward Skipping toStartOfMicrosecond Checking toInt8OrNull('foo') Executing query select toInt8OrNull('foo') on backward Executing query select 20 on node Executing query select toInt8OrNull('foo') on upstream OK toInt8OrNull Checking regionToPopulation('foo') Executing query select regionToPopulation('foo') on backward Skipping regionToPopulation Checking toStartOfISOYear('foo') Executing query select toStartOfISOYear('foo') on backward Skipping toStartOfISOYear Checking murmurHash2_32('foo') Executing query select murmurHash2_32('foo') on backward Executing query select murmurHash2_32('foo') on upstream OK murmurHash2_32 Checking toStartOfTenMinutes('foo') Executing query select toStartOfTenMinutes('foo') on backward Skipping toStartOfTenMinutes Checking arrayEnumerateUniq('foo') Executing query select arrayEnumerateUniq('foo') on backward Skipping arrayEnumerateUniq Checking bitmaskToArray('foo') Executing query select bitmaskToArray('foo') on backward Executing query select 20 on node Skipping bitmaskToArray Checking min2('foo') Executing query select min2('foo') on backward Skipping min2 Checking length('foo') Executing query select length('foo') on backward Executing query select length('foo') on upstream OK length Checking toHour('foo') Executing query select toHour('foo') on backward Skipping toHour Checking toStartOfFifteenMinutes('foo') Executing query select toStartOfFifteenMinutes('foo') on backward Skipping toStartOfFifteenMinutes Checking dictHas('foo') Executing query select dictHas('foo') on backward run container_id:roottestcancelfreeze-gw1-node-1 detach:False nothrow:False cmd: ['bash', '-c', 'rm -r /var/lib/clickhouse/shadow/'] Command:[docker exec roottestcancelfreeze-gw1-node-1 bash -c rm -r /var/lib/clickhouse/shadow/] Skipping dictHas Checking enabledProfiles('foo') Executing query select enabledProfiles('foo') on backward Skipping enabledProfiles Checking toYYYYMMDD('foo') Executing query select toYYYYMMDD('foo') on backward Executing query select 20 on node Command:[docker compose --env-file /ClickHouse/tests/integration/test_cancel_freeze/_instances-0-gw1/.env --project-name roottestcancelfreeze-gw1 --file /ClickHouse/tests/integration/test_cancel_freeze/_instances-0-gw1/node/docker-compose.yml stop --timeout 20] Skipping toYYYYMMDD Checking toUInt256OrNull('foo') Executing query select toUInt256OrNull('foo') on backward Executing query select toUInt256OrNull('foo') on upstream OK toUInt256OrNull Checking arrayFirstOrNull('foo') Executing query select arrayFirstOrNull('foo') on backward Skipping arrayFirstOrNull Checking isNullable('foo') Executing query select isNullable('foo') on backward Executing query select isNullable('foo') on upstream OK isNullable Checking regexpExtract('foo') Executing query select regexpExtract('foo') on backward Skipping regexpExtract Checking cutToFirstSignificantSubdomainCustomRFC('foo') Executing query select cutToFirstSignificantSubdomainCustomRFC('foo') on backward Executing query select 20 on node Skipping cutToFirstSignificantSubdomainCustomRFC Checking multiMatchAnyIndex('foo') Executing query select multiMatchAnyIndex('foo') on backward Skipping multiMatchAnyIndex Checking toFloat64OrDefault('foo') Executing query select toFloat64OrDefault('foo') on backward Executing query select toFloat64OrDefault('foo') on upstream OK toFloat64OrDefault Checking arrayLast('foo') Executing query select arrayLast('foo') on backward Skipping arrayLast Checking bitPositionsToArray('foo') Executing query select bitPositionsToArray('foo') on backward Skipping bitPositionsToArray Checking xor('foo') Executing query select xor('foo') on backward Skipping xor Checking toInt256('foo') Executing query select toInt256('foo') on backward Skipping toInt256 Checking sign('foo') Executing query select sign('foo') on backward run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Skipping sign Checking mapKeys('foo') Executing query select mapKeys('foo') on backward Current start attempt failed. Will kill 6590 just in case. run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:True cmd: ['bash', '-c', 'kill -9 6590'] Command:[docker exec -u root roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c kill -9 6590] Skipping mapKeys Checking tupleElement('foo') Executing query select tupleElement('foo') on backward Stderr:bash: line 0: kill: (6590) - No such process Exitcode:1 Skipping tupleElement Checking addYears('foo') Executing query select addYears('foo') on backward Skipping addYears Checking validateNestedArraySizes('foo') Executing query select validateNestedArraySizes('foo') on backward Skipping validateNestedArraySizes Checking addHours('foo') Executing query select addHours('foo') on backward Skipping addHours Checking getOSKernelVersion('foo') Executing query select getOSKernelVersion('foo') on backward Skipping getOSKernelVersion Checking reinterpretAsUInt32('foo') Executing query select reinterpretAsUInt32('foo') on backward Executing query select reinterpretAsUInt32('foo') on upstream OK reinterpretAsUInt32 Checking greaterOrEquals('foo') Executing query select greaterOrEquals('foo') on backward run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Skipping greaterOrEquals Checking upper('foo') Executing query select upper('foo') on backward No clickhouse process running. Start new one. http://localhost:None "POST /v1.46/containers/roottestbackwardcompatibilityconvertordinary-gw5-node-1/exec HTTP/1.1" 201 74 http://localhost:None "POST /v1.46/exec/d3899a57f71319669d547b489f7b8aec3f017253530d65d42cbe91389805c101/start HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/exec/d3899a57f71319669d547b489f7b8aec3f017253530d65d42cbe91389805c101/json HTTP/1.1" 200 587 Executing query select upper('foo') on upstream OK upper Checking bitmapTransform('foo') Executing query select bitmapTransform('foo') on backward Skipping bitmapTransform Checking arrayShuffle('foo') Executing query select arrayShuffle('foo') on backward Skipping arrayShuffle Checking toISOWeek('foo') Executing query select toISOWeek('foo') on backward Skipping toISOWeek Checking reinterpretAsInt256('foo') Executing query select reinterpretAsInt256('foo') on backward Executing query select reinterpretAsInt256('foo') on upstream OK reinterpretAsInt256 Checking hasAny('foo') Executing query select hasAny('foo') on backward Skipping hasAny Checking toIntervalMicrosecond('foo') Executing query select toIntervalMicrosecond('foo') on backward Skipping toIntervalMicrosecond Checking L1Normalize('foo') Executing query select L1Normalize('foo') on backward Skipping L1Normalize Checking multiply('foo') Executing query select multiply('foo') on backward Skipping multiply Checking toIPv6('foo') Executing query select toIPv6('foo') on backward Skipping toIPv6 Checking toDateTime32('foo') Executing query select toDateTime32('foo') on backward Skipping toDateTime32 Checking arraySplit('foo') Executing query select arraySplit('foo') on backward Skipping arraySplit Checking caseWithExpr('foo') Executing query select caseWithExpr('foo') on backward Skipping caseWithExpr Checking s2CellsIntersect('foo') Executing query select s2CellsIntersect('foo') on backward run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Skipping s2CellsIntersect Checking decodeURLFormComponent('foo') Executing query select decodeURLFormComponent('foo') on backward Stdout:7353 Clickhouse process running. run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:7353 Executing query select 20 on node Executing query select decodeURLFormComponent('foo') on upstream OK decodeURLFormComponent Checking L2Distance('foo') Executing query select L2Distance('foo') on backward Skipping L2Distance Checking proportionsZTest('foo') Executing query select proportionsZTest('foo') on backward Skipping proportionsZTest Checking arrayZip('foo') Executing query select arrayZip('foo') on backward Connection dropped: socket connection error: No route to host Skipping arrayZip Checking displayName('foo') Executing query select displayName('foo') on backward Stderr: Container roottestcancelfreeze-gw1-node-1 Stopping Stderr: Container roottestcancelfreeze-gw1-node-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_cancel_freeze/_instances-0-gw1/node/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_cancel_freeze/_instances-0-gw1/node/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_cancel_freeze/_instances-0-gw1/.env --project-name roottestcancelfreeze-gw1 --file /ClickHouse/tests/integration/test_cancel_freeze/_instances-0-gw1/node/docker-compose.yml down --volumes] Skipping displayName Checking L1Distance('foo') Executing query select L1Distance('foo') on backward Executing query select 20 on node Skipping L1Distance Checking emptyArrayFloat64('foo') Executing query select emptyArrayFloat64('foo') on backward Skipping emptyArrayFloat64 Checking LinfNorm('foo') Executing query select LinfNorm('foo') on backward Skipping LinfNorm Checking less('foo') Executing query select less('foo') on backward Skipping less Checking emptyArrayInt8('foo') Executing query select emptyArrayInt8('foo') on backward Skipping emptyArrayInt8 Checking L2SquaredNorm('foo') Executing query select L2SquaredNorm('foo') on backward Skipping L2SquaredNorm Checking formatReadableDecimalSize('foo') Executing query select formatReadableDecimalSize('foo') on backward Skipping formatReadableDecimalSize Checking lowerUTF8('foo') Executing query select lowerUTF8('foo') on backward Stderr: Container roottestcancelfreeze-gw1-node-1 Stopping Stderr: Container roottestcancelfreeze-gw1-node-1 Stopped Stderr: Container roottestcancelfreeze-gw1-node-1 Removing Stderr: Container roottestcancelfreeze-gw1-node-1 Removed Stderr: Network roottestcancelfreeze-gw1_default Removing Stderr: Network roottestcancelfreeze-gw1_default Removed Cleanup called Docker networks for project roottestcancelfreeze-gw1 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestcancelfreeze-gw1 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestcancelfreeze-gw1 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestcancelfreeze-gw1-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestcancelfreeze-gw1 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Executing query select lowerUTF8('foo') on upstream Executing query select 20 on node Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:1 Volumes pruned: 1 OK lowerUTF8 Checking parseDateTime64BestEffortOrNull('foo') Executing query select parseDateTime64BestEffortOrNull('foo') on backward Executing query select parseDateTime64BestEffortOrNull('foo') on upstream OK parseDateTime64BestEffortOrNull Checking domainRFC('foo') Executing query select domainRFC('foo') on backward Executing query select domainRFC('foo') on upstream OK domainRFC Checking coalesce('foo') Executing query select coalesce('foo') on backward Executing query select coalesce('foo') on upstream OK coalesce Checking basename('foo') Executing query select basename('foo') on backward Executing query select basename('foo') on upstream Executing query select 20 on node OK basename Checking ULIDStringToDateTime('foo') Executing query select ULIDStringToDateTime('foo') on backward Skipping ULIDStringToDateTime Checking JSONType('foo') Executing query select JSONType('foo') on backward Executing query select JSONType('foo') on upstream OK JSONType Checking addNanoseconds('foo') Executing query select addNanoseconds('foo') on backward Skipping addNanoseconds Checking isConstant('foo') Executing query select isConstant('foo') on backward Executing query select isConstant('foo') on upstream OK isConstant Checking toRelativeHourNum('foo') Executing query select toRelativeHourNum('foo') on backward Skipping toRelativeHourNum Checking hasColumnInTable('foo') Executing query select hasColumnInTable('foo') on backward Skipping hasColumnInTable Checking arrayProduct('foo') Executing query select arrayProduct('foo') on backward Skipping arrayProduct Checking polygonsIntersectionSpherical('foo') Executing query select polygonsIntersectionSpherical('foo') on backward Executing query select 20 on node Skipping polygonsIntersectionSpherical Checking TimeDiff('foo') Executing query select TimeDiff('foo') on backward Skipping TimeDiff Checking polygonAreaCartesian('foo') Executing query select polygonAreaCartesian('foo') on backward Skipping polygonAreaCartesian Checking arrayUniq('foo') Executing query select arrayUniq('foo') on backward Skipping arrayUniq Checking toIntervalYear('foo') Executing query select toIntervalYear('foo') on backward Skipping toIntervalYear Checking addInterval('foo') Executing query select addInterval('foo') on backward Skipping addInterval Checking bitRotateLeft('foo') Executing query select bitRotateLeft('foo') on backward Skipping bitRotateLeft Checking toUInt16OrZero('foo') Executing query select toUInt16OrZero('foo') on backward Executing query select toUInt16OrZero('foo') on upstream OK toUInt16OrZero Checking emptyArrayDateTime('foo') Executing query select emptyArrayDateTime('foo') on backward Executing query select 20 on node Skipping emptyArrayDateTime Checking dictGetFloat32('foo') Executing query select dictGetFloat32('foo') on backward Skipping dictGetFloat32 Checking h3PointDistRads('foo') Executing query select h3PointDistRads('foo') on backward Skipping h3PointDistRads Checking polygonsUnionSpherical('foo') Executing query select polygonsUnionSpherical('foo') on backward Skipping polygonsUnionSpherical Checking arraySum('foo') Executing query select arraySum('foo') on backward Skipping arraySum Checking dictGetDateOrDefault('foo') Executing query select dictGetDateOrDefault('foo') on backward Skipping dictGetDateOrDefault Checking wordShingleMinHashCaseInsensitive('foo') Executing query select wordShingleMinHashCaseInsensitive('foo') on backward Executing query select wordShingleMinHashCaseInsensitive('foo') on upstream OK wordShingleMinHashCaseInsensitive Checking toInt128OrDefault('foo') Executing query select toInt128OrDefault('foo') on backward Executing query select 20 on node Executing query select toInt128OrDefault('foo') on upstream OK toInt128OrDefault Checking addTupleOfIntervals('foo') Executing query select addTupleOfIntervals('foo') on backward Skipping addTupleOfIntervals Checking tupleMultiply('foo') Executing query select tupleMultiply('foo') on backward Skipping tupleMultiply Checking cutToFirstSignificantSubdomainCustom('foo') Executing query select cutToFirstSignificantSubdomainCustom('foo') on backward Skipping cutToFirstSignificantSubdomainCustom Checking dictGetInt16OrDefault('foo') Executing query select dictGetInt16OrDefault('foo') on backward Skipping dictGetInt16OrDefault Checking simpleJSONExtractInt('foo') Executing query select simpleJSONExtractInt('foo') on backward Skipping simpleJSONExtractInt Checking addMonths('foo') Executing query select addMonths('foo') on backward Skipping addMonths Checking leftUTF8('foo') Executing query select leftUTF8('foo') on backward Skipping leftUTF8 Checking topLevelDomainRFC('foo') Executing query select topLevelDomainRFC('foo') on backward Executing query select 20 on node Executing query select topLevelDomainRFC('foo') on upstream OK topLevelDomainRFC Checking has('foo') Executing query select has('foo') on backward Skipping has Checking appendTrailingCharIfAbsent('foo') Executing query select appendTrailingCharIfAbsent('foo') on backward Skipping appendTrailingCharIfAbsent Checking simpleJSONExtractRaw('foo') Executing query select simpleJSONExtractRaw('foo') on backward Skipping simpleJSONExtractRaw Checking arrayJoin('foo') Executing query select arrayJoin('foo') on backward Skipping arrayJoin Checking simpleJSONHas('foo') Executing query select simpleJSONHas('foo') on backward Skipping simpleJSONHas Checking concatWithSeparatorAssumeInjective('foo') Executing query select concatWithSeparatorAssumeInjective('foo') on backward Executing query select concatWithSeparatorAssumeInjective('foo') on upstream OK concatWithSeparatorAssumeInjective Checking toDateTime64OrDefault('foo') Executing query select toDateTime64OrDefault('foo') on backward Executing query select 20 on node Skipping toDateTime64OrDefault Checking leftPad('foo') Executing query select leftPad('foo') on backward Skipping leftPad Checking fromUnixTimestamp64Micro('foo') Executing query select fromUnixTimestamp64Micro('foo') on backward Skipping fromUnixTimestamp64Micro Checking range('foo') Executing query select range('foo') on backward Skipping range Checking arrayPartialSort('foo') Executing query select arrayPartialSort('foo') on backward Skipping arrayPartialSort Checking toJSONString('foo') Executing query select toJSONString('foo') on backward Executing query select toJSONString('foo') on upstream OK toJSONString Checking __getScalar('foo') Executing query select __getScalar('foo') on backward Skipping __getScalar Checking h3GetFaces('foo') Executing query select h3GetFaces('foo') on backward Skipping h3GetFaces Checking toUInt32OrNull('foo') Executing query select toUInt32OrNull('foo') on backward Executing query select 20 on node Executing query select toUInt32OrNull('foo') on upstream OK toUInt32OrNull Checking accurateCast('foo') Executing query select accurateCast('foo') on backward Skipping accurateCast Checking toInt64('foo') Executing query select toInt64('foo') on backward Skipping toInt64 Checking wkt('foo') Executing query select wkt('foo') on backward Skipping wkt Checking domainWithoutWWW('foo') Executing query select domainWithoutWWW('foo') on backward Executing query select domainWithoutWWW('foo') on upstream OK domainWithoutWWW Checking arrayPushFront('foo') Executing query select arrayPushFront('foo') on backward Skipping arrayPushFront Checking erfc('foo') Executing query select erfc('foo') on backward Skipping erfc Checking cutToFirstSignificantSubdomainCustomWithWWW('foo') Executing query select cutToFirstSignificantSubdomainCustomWithWWW('foo') on backward run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Skipping cutToFirstSignificantSubdomainCustomWithWWW Checking arraySort('foo') Executing query select arraySort('foo') on backward Current start attempt failed. Will kill 7353 just in case. run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:True cmd: ['bash', '-c', 'kill -9 7353'] Command:[docker exec -u root roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c kill -9 7353] Skipping arraySort Checking firstSignificantSubdomainCustomRFC('foo') Executing query select firstSignificantSubdomainCustomRFC('foo') on backward Stderr:bash: line 0: kill: (7353) - No such process Exitcode:1 Skipping firstSignificantSubdomainCustomRFC Checking tupleMinus('foo') Executing query select tupleMinus('foo') on backward Skipping tupleMinus Checking uniqThetaUnion('foo') Executing query select uniqThetaUnion('foo') on backward Skipping uniqThetaUnion Checking gccMurmurHash('foo') Executing query select gccMurmurHash('foo') on backward Executing query select gccMurmurHash('foo') on upstream OK gccMurmurHash Checking polygonsSymDifferenceCartesian('foo') Executing query select polygonsSymDifferenceCartesian('foo') on backward Skipping polygonsSymDifferenceCartesian Checking LinfDistance('foo') Executing query select LinfDistance('foo') on backward Skipping LinfDistance Checking readWKTPoint('foo') Executing query select readWKTPoint('foo') on backward run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', 'rm /var/lib/clickhouse/flags/convert_ordinary_to_atomic'] Command:[docker exec roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c rm /var/lib/clickhouse/flags/convert_ordinary_to_atomic] Skipping readWKTPoint Checking randChiSquared('foo') Executing query select randChiSquared('foo') on backward run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Skipping randChiSquared Checking IPv4StringToNumOrNull('foo') Executing query select IPv4StringToNumOrNull('foo') on backward No clickhouse process running. Start new one. http://localhost:None "POST /v1.46/containers/roottestbackwardcompatibilityconvertordinary-gw5-node-1/exec HTTP/1.1" 201 74 http://localhost:None "POST /v1.46/exec/7f469045b4612570771a9a2edb15ac94a3640054313042b9d78d569cf60209d7/start HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/exec/7f469045b4612570771a9a2edb15ac94a3640054313042b9d78d569cf60209d7/json HTTP/1.1" 200 587 Executing query select IPv4StringToNumOrNull('foo') on upstream OK IPv4StringToNumOrNull Checking formatReadableQuantity('foo') Executing query select formatReadableQuantity('foo') on backward Skipping formatReadableQuantity Checking arrayPopBack('foo') Executing query select arrayPopBack('foo') on backward Skipping arrayPopBack Checking arrayLastIndex('foo') Executing query select arrayLastIndex('foo') on backward Skipping arrayLastIndex Checking toUnixTimestamp64Nano('foo') Executing query select toUnixTimestamp64Nano('foo') on backward Skipping toUnixTimestamp64Nano Checking subtractMicroseconds('foo') Executing query select subtractMicroseconds('foo') on backward Skipping subtractMicroseconds Checking decodeXMLComponent('foo') Executing query select decodeXMLComponent('foo') on backward Executing query select decodeXMLComponent('foo') on upstream OK decodeXMLComponent Checking svg('foo') Executing query select svg('foo') on backward Skipping svg Checking getMacro('foo') Executing query select getMacro('foo') on backward Skipping getMacro Checking multiIf('foo') Executing query select multiIf('foo') on backward Skipping multiIf Checking runningAccumulate('foo') Executing query select runningAccumulate('foo') on backward Skipping runningAccumulate Checking SHA384('foo') Executing query select SHA384('foo') on backward Executing query select SHA384('foo') on upstream OK SHA384 Checking murmurHash3_128('foo') Executing query select murmurHash3_128('foo') on backward Executing query select murmurHash3_128('foo') on upstream run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] OK murmurHash3_128 Checking ngramSearchCaseInsensitive('foo') Executing query select ngramSearchCaseInsensitive('foo') on backward Skipping ngramSearchCaseInsensitive Checking h3ToCenterChild('foo') Executing query select h3ToCenterChild('foo') on backward Stdout:8122 Clickhouse process running. run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Skipping h3ToCenterChild Checking notIn('foo') Executing query select notIn('foo') on backward Stdout:8122 Executing query select 20 on node Skipping notIn Checking URLHierarchy('foo') Executing query select URLHierarchy('foo') on backward Executing query SELECT engine FROM system.databases where name='ordinary' on node Executing query select URLHierarchy('foo') on upstream Executing query ATTACH TABLE ordinary.detached on node OK URLHierarchy Checking ngramSimHashUTF8('foo') Executing query select ngramSimHashUTF8('foo') on backward run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', 'touch /var/lib/clickhouse/flags/convert_ordinary_to_atomic'] Command:[docker exec roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c touch /var/lib/clickhouse/flags/convert_ordinary_to_atomic] run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] Command:[docker exec -u root roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c ps -C clickhouse] Executing query select ngramSimHashUTF8('foo') on upstream Stdout: PID TTY TIME CMD Stdout: 8122 ? 00:00:01 clickhouse run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] Command:[docker exec -u root roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c pkill clickhouse] OK ngramSimHashUTF8 Checking subtractTupleOfIntervals('foo') Executing query select subtractTupleOfIntervals('foo') on backward run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:8122 Skipping subtractTupleOfIntervals Checking toUInt16OrDefault('foo') Executing query select toUInt16OrDefault('foo') on backward Executing query select toUInt16OrDefault('foo') on upstream OK toUInt16OrDefault Checking s2ToGeo('foo') Executing query select s2ToGeo('foo') on backward Skipping s2ToGeo Checking sinh('foo') Executing query select sinh('foo') on backward Skipping sinh Checking tupleDivide('foo') Executing query select tupleDivide('foo') on backward Skipping tupleDivide Checking JSONExtractInt('foo') Executing query select JSONExtractInt('foo') on backward Executing query select JSONExtractInt('foo') on upstream OK JSONExtractInt Checking subtractMonths('foo') Executing query select subtractMonths('foo') on backward Skipping subtractMonths Checking subtractMilliseconds('foo') Executing query select subtractMilliseconds('foo') on backward Skipping subtractMilliseconds Checking arrayPopFront('foo') Executing query select arrayPopFront('foo') on backward Skipping arrayPopFront Checking dictGetFloat32OrDefault('foo') Executing query select dictGetFloat32OrDefault('foo') on backward Skipping dictGetFloat32OrDefault Checking s2RectAdd('foo') Executing query select s2RectAdd('foo') on backward Skipping s2RectAdd Checking toColumnTypeName('foo') Executing query select toColumnTypeName('foo') on backward Executing query select toColumnTypeName('foo') on upstream OK toColumnTypeName Checking arrayMin('foo') Executing query select arrayMin('foo') on backward Skipping arrayMin Checking sin('foo') Executing query select sin('foo') on backward run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:8122 Skipping sin Checking multiFuzzyMatchAllIndices('foo') Executing query select multiFuzzyMatchAllIndices('foo') on backward Skipping multiFuzzyMatchAllIndices Checking toStartOfSecond('foo') Executing query select toStartOfSecond('foo') on backward Skipping toStartOfSecond Checking dictGet('foo') Executing query select dictGet('foo') on backward Skipping dictGet Checking cos('foo') Executing query select cos('foo') on backward Skipping cos Checking toDecimal128OrDefault('foo') Executing query select toDecimal128OrDefault('foo') on backward Skipping toDecimal128OrDefault Checking arrayCount('foo') Executing query select arrayCount('foo') on backward Skipping arrayCount Checking hex('foo') Executing query select hex('foo') on backward Executing query select hex('foo') on upstream OK hex Checking h3ToGeo('foo') Executing query select h3ToGeo('foo') on backward Skipping h3ToGeo Checking toDecimal256OrNull('foo') Executing query select toDecimal256OrNull('foo') on backward Skipping toDecimal256OrNull Checking multiMatchAny('foo') Executing query select multiMatchAny('foo') on backward Skipping multiMatchAny Checking arrayEnumerateDense('foo') Executing query select arrayEnumerateDense('foo') on backward Skipping arrayEnumerateDense Checking or('foo') Executing query select or('foo') on backward Skipping or Checking arrayExists('foo') Executing query select arrayExists('foo') on backward Skipping arrayExists Checking L2Norm('foo') Executing query select L2Norm('foo') on backward Skipping L2Norm Checking h3GetDestinationIndexFromUnidirectionalEdge('foo') Executing query select h3GetDestinationIndexFromUnidirectionalEdge('foo') on backward run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Skipping h3GetDestinationIndexFromUnidirectionalEdge Checking parseDateTimeBestEffortUSOrZero('foo') Executing query select parseDateTimeBestEffortUSOrZero('foo') on backward Stdout:8122 Executing query select parseDateTimeBestEffortUSOrZero('foo') on upstream OK parseDateTimeBestEffortUSOrZero Checking toInt32OrNull('foo') Executing query select toInt32OrNull('foo') on backward Executing query select toInt32OrNull('foo') on upstream OK toInt32OrNull Checking array('foo') Executing query select array('foo') on backward Executing query select array('foo') on upstream OK array Checking toLowCardinality('foo') Executing query select toLowCardinality('foo') on backward Executing query select toLowCardinality('foo') on upstream OK toLowCardinality Checking h3ToString('foo') Executing query select h3ToString('foo') on backward Skipping h3ToString Checking LpNorm('foo') Executing query select LpNorm('foo') on backward Skipping LpNorm Checking port('foo') Executing query select port('foo') on backward Executing query select port('foo') on upstream OK port Checking wordShingleMinHashArgUTF8('foo') Executing query select wordShingleMinHashArgUTF8('foo') on backward Executing query select wordShingleMinHashArgUTF8('foo') on upstream run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:8122 OK wordShingleMinHashArgUTF8 Checking now('foo') Executing query select now('foo') on backward Skipping now Checking timezone('foo') Executing query select timezone('foo') on backward Skipping timezone Checking arrayMap('foo') Executing query select arrayMap('foo') on backward Skipping arrayMap Checking toInt32OrDefault('foo') Executing query select toInt32OrDefault('foo') on backward Executing query select toInt32OrDefault('foo') on upstream OK toInt32OrDefault Checking mapAdd('foo') Executing query select mapAdd('foo') on backward Skipping mapAdd Checking countSubstringsCaseInsensitive('foo') Executing query select countSubstringsCaseInsensitive('foo') on backward Skipping countSubstringsCaseInsensitive Checking uniqThetaNot('foo') Executing query select uniqThetaNot('foo') on backward Skipping uniqThetaNot Checking indexHint('foo') Executing query select indexHint('foo') on backward Executing query select indexHint('foo') on upstream OK indexHint Checking toMinute('foo') Executing query select toMinute('foo') on backward Skipping toMinute Checking tupleToNameValuePairs('foo') Executing query select tupleToNameValuePairs('foo') on backward Skipping tupleToNameValuePairs Checking JSONExtractUInt('foo') Executing query select JSONExtractUInt('foo') on backward Executing query select JSONExtractUInt('foo') on upstream OK JSONExtractUInt Checking arrayReduce('foo') Executing query select arrayReduce('foo') on backward Skipping arrayReduce Checking addDays('foo') Executing query select addDays('foo') on backward Skipping addDays Checking portRFC('foo') Executing query select portRFC('foo') on backward run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select portRFC('foo') on upstream No clickhouse process running. Start new one. http://localhost:None "POST /v1.46/containers/roottestbackwardcompatibilityconvertordinary-gw5-node-1/exec HTTP/1.1" 201 74 http://localhost:None "POST /v1.46/exec/b8bb13bea993e7335d620b7f0cca60921a2c53fc1d235b3cca66e4df840bb18b/start HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/exec/b8bb13bea993e7335d620b7f0cca60921a2c53fc1d235b3cca66e4df840bb18b/json HTTP/1.1" 200 587 OK portRFC Checking L1Norm('foo') Executing query select L1Norm('foo') on backward Skipping L1Norm Checking log1p('foo') Executing query select log1p('foo') on backward Skipping log1p Checking trunc('foo') Executing query select trunc('foo') on backward Skipping trunc Checking base64Encode('foo') Executing query select base64Encode('foo') on backward Executing query select base64Encode('foo') on upstream OK base64Encode Checking IPv6NumToString('foo') Executing query select IPv6NumToString('foo') on backward Skipping IPv6NumToString Checking simpleJSONExtractUInt('foo') Executing query select simpleJSONExtractUInt('foo') on backward Skipping simpleJSONExtractUInt Checking toTimezone('foo') Executing query select toTimezone('foo') on backward Skipping toTimezone Checking currentDatabase('foo') Executing query select currentDatabase('foo') on backward Skipping currentDatabase Checking toMonth('foo') Executing query select toMonth('foo') on backward Skipping toMonth Checking positiveModulo('foo') Executing query select positiveModulo('foo') on backward Skipping positiveModulo Checking positiveModulo('foo') Executing query select positiveModulo('foo') on backward Skipping positiveModulo Checking dateTrunc('foo') Executing query select dateTrunc('foo') on backward Skipping dateTrunc Checking toStartOfFiveMinutes('foo') Executing query select toStartOfFiveMinutes('foo') on backward Skipping toStartOfFiveMinutes Checking toYear('foo') Executing query select toYear('foo') on backward Skipping toYear Checking concatWithSeparator('foo') Executing query select concatWithSeparator('foo') on backward run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select concatWithSeparator('foo') on upstream Stdout:8940 Clickhouse process running. run container_id:roottestbackwardcompatibilityconvertordinary-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackwardcompatibilityconvertordinary-gw5-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:8940 Executing query select 20 on node OK concatWithSeparator Checking fromUnixTimestamp('foo') Executing query select fromUnixTimestamp('foo') on backward Skipping fromUnixTimestamp Checking kostikConsistentHash('foo') Executing query select kostikConsistentHash('foo') on backward Executing query SELECT name FROM system.databases WHERE engine='Atomic' ORDER BY name on node Skipping kostikConsistentHash Checking lengthUTF8('foo') Executing query select lengthUTF8('foo') on backward Executing query SELECT engine FROM system.databases WHERE name IN ('mem', 'lazy') ORDER BY name on node Executing query select lengthUTF8('foo') on upstream Executing query SHOW TABLES FROM `.o r d i n a r y.` on node OK lengthUTF8 Checking rightPad('foo') Executing query select rightPad('foo') on backward Skipping rightPad Checking UTCTimestamp('foo') Executing query select UTCTimestamp('foo') on backward Executing query SHOW TABLES FROM lazy on node Skipping UTCTimestamp Checking currentUser('foo') Executing query select currentUser('foo') on backward Executing query SHOW TABLES FROM ordinary NOT LIKE '%inner%' on node Skipping currentUser Checking lower('foo') Executing query select lower('foo') on backward Executing query SHOW TABLES FROM other NOT LIKE '%inner%' on node Executing query select lower('foo') on upstream Executing query SHOW TABLES FROM atomic NOT LIKE '%inner%' on node OK lower Checking match('foo') Executing query select match('foo') on backward Executing query SELECT count(), sum(n) FROM ordinary.merge on node Skipping match Checking toSecond('foo') Executing query select toSecond('foo') on backward Skipping toSecond Checking formatDateTime('foo') Executing query select formatDateTime('foo') on backward Skipping formatDateTime Checking mapFromArrays('foo') Executing query select mapFromArrays('foo') on backward Executing query SELECT count(), sum(n) FROM other.merge on node Skipping mapFromArrays Checking pow('foo') Executing query select pow('foo') on backward Executing query INSERT INTO ordinary.mt1 (n) VALUES (24) on node Skipping pow Checking simpleJSONExtractInt('foo') Executing query select simpleJSONExtractInt('foo') on backward Executing query INSERT INTO ordinary.mt2 (n) VALUES (24) on node Skipping simpleJSONExtractInt Checking toYearWeek('foo') Executing query select toYearWeek('foo') on backward Executing query INSERT INTO ordinary.rmt1 (n) VALUES (24) on node Skipping toYearWeek Checking replaceRegexpAll('foo') Executing query select replaceRegexpAll('foo') on backward Executing query INSERT INTO ordinary.rmt2 (n) VALUES (24) on node Skipping replaceRegexpAll Checking timezoneOf('foo') Executing query select timezoneOf('foo') on backward Executing query INSERT INTO ordinary.mv1 (n) VALUES (24) on node Skipping timezoneOf Checking LpNormalize('foo') Executing query select LpNormalize('foo') on backward Executing query INSERT INTO ordinary.mv2 (n) VALUES (24) on node Skipping LpNormalize Checking toWeek('foo') Executing query select toWeek('foo') on backward Executing query INSERT INTO ordinary.detached (n) VALUES (24) on node Skipping toWeek Checking toDayOfYear('foo') Executing query select toDayOfYear('foo') on backward Executing query INSERT INTO other.mt1 (n) VALUES (15) on node Skipping toDayOfYear Checking ceil('foo') Executing query select ceil('foo') on backward Executing query INSERT INTO other.mt2 (n) VALUES (15) on node Skipping ceil Checking toHour('foo') Executing query select toHour('foo') on backward Executing query INSERT INTO other.rmt1 (n) VALUES (15) on node Skipping toHour Checking extractAllGroupsVertical('foo') Executing query select extractAllGroupsVertical('foo') on backward Executing query INSERT INTO other.rmt2 (n) VALUES (15) on node Skipping extractAllGroupsVertical Checking tupleMinus('foo') Executing query select tupleMinus('foo') on backward Executing query INSERT INTO other.mv1 (n) VALUES (15) on node Skipping tupleMinus Checking toLastDayOfMonth('foo') Executing query select toLastDayOfMonth('foo') on backward Skipping toLastDayOfMonth Checking simpleJSONExtractRaw('foo') Executing query select simpleJSONExtractRaw('foo') on backward Executing query INSERT INTO other.mv2 (n) VALUES (15) on node Skipping simpleJSONExtractRaw Checking arrayFlatten('foo') Executing query select arrayFlatten('foo') on backward Executing query INSERT INTO other.detached (n) VALUES (15) on node Skipping arrayFlatten Checking replaceAll('foo') Executing query select replaceAll('foo') on backward Executing query INSERT INTO atomic.mt1 (n) VALUES (18) on node Skipping replaceAll Checking toQuarter('foo') Executing query select toQuarter('foo') on backward Executing query INSERT INTO atomic.mt2 (n) VALUES (18) on node Skipping toQuarter Checking positionCaseInsensitive('foo') Executing query select positionCaseInsensitive('foo') on backward Executing query INSERT INTO atomic.rmt1 (n) VALUES (18) on node Skipping positionCaseInsensitive Checking toDayOfWeek('foo') Executing query select toDayOfWeek('foo') on backward Executing query INSERT INTO atomic.rmt2 (n) VALUES (18) on node Skipping toDayOfWeek Checking LinfNormalize('foo') Executing query select LinfNormalize('foo') on backward Executing query INSERT INTO atomic.mv1 (n) VALUES (18) on node Skipping LinfNormalize Checking timezone('foo') Executing query select timezone('foo') on backward Executing query INSERT INTO atomic.mv2 (n) VALUES (18) on node Skipping timezone Checking toMinute('foo') Executing query select toMinute('foo') on backward Executing query INSERT INTO atomic.detached (n) VALUES (18) on node Skipping toMinute Checking initialQueryID('foo') Executing query select initialQueryID('foo') on backward Executing query SELECT count(), sum(n) FROM ordinary.merge on node Skipping initialQueryID Checking widthBucket('foo') Executing query select widthBucket('foo') on backward Executing query SELECT count(), sum(n) FROM ordinary.detached on node Skipping widthBucket Checking leftPad('foo') Executing query select leftPad('foo') on backward Skipping leftPad Checking L1Normalize('foo') Executing query select L1Normalize('foo') on backward Executing query SELECT count(), sum(n) FROM other.merge on node Skipping L1Normalize Checking lengthUTF8('foo') Executing query select lengthUTF8('foo') on backward Executing query SELECT count(), sum(n) FROM other.detached on node Executing query select lengthUTF8('foo') on upstream Executing query SELECT count(), sum(n) FROM atomic.merge on node OK lengthUTF8 Checking simpleJSONExtractFloat('foo') Executing query select simpleJSONExtractFloat('foo') on backward Executing query SELECT count(), sum(n) FROM atomic.detached on node Skipping simpleJSONExtractFloat Checking LinfDistance('foo') Executing query select LinfDistance('foo') on backward Command:[docker compose --env-file /ClickHouse/tests/integration/test_backward_compatibility/_instances-convert_ordinary-0-gw5/.env --project-name roottestbackwardcompatibilityconvertordinary-gw5 --file /ClickHouse/tests/integration/test_backward_compatibility/_instances-convert_ordinary-0-gw5/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml stop --timeout 20] [gw5] PASSED test_backward_compatibility/test_convert_ordinary.py::test_convert_ordinary_to_atomic Skipping LinfDistance Checking toDayOfMonth('foo') Executing query select toDayOfMonth('foo') on backward Skipping toDayOfMonth Checking alphaTokens('foo') Executing query select alphaTokens('foo') on backward Executing query select alphaTokens('foo') on upstream OK alphaTokens Checking log('foo') Executing query select log('foo') on backward Skipping log Checking LpDistance('foo') Executing query select LpDistance('foo') on backward Skipping LpDistance Checking IPv6StringToNum('foo') Executing query select IPv6StringToNum('foo') on backward Skipping IPv6StringToNum Checking simpleJSONExtractString('foo') Executing query select simpleJSONExtractString('foo') on backward Skipping simpleJSONExtractString Checking toDayOfMonth('foo') Executing query select toDayOfMonth('foo') on backward Skipping toDayOfMonth Checking LpNorm('foo') Executing query select LpNorm('foo') on backward Skipping LpNorm Checking IPv4StringToNum('foo') Executing query select IPv4StringToNum('foo') on backward Skipping IPv4StringToNum Checking L1Distance('foo') Executing query select L1Distance('foo') on backward Skipping L1Distance Checking upper('foo') Executing query select upper('foo') on backward Executing query select upper('foo') on upstream OK upper Checking JSONArrayLength('foo') Executing query select JSONArrayLength('foo') on backward Executing query select JSONArrayLength('foo') on upstream OK JSONArrayLength Checking LinfNorm('foo') Executing query select LinfNorm('foo') on backward Skipping LinfNorm Checking L2SquaredDistance('foo') Executing query select L2SquaredDistance('foo') on backward Skipping L2SquaredDistance Checking queryID('foo') Executing query select queryID('foo') on backward Skipping queryID Checking L2SquaredNorm('foo') Executing query select L2SquaredNorm('foo') on backward Stderr: Container roottestbackwardcompatibilityconvertordinary-gw5-node-1 Stopping Stderr: Container roottestbackwardcompatibilityconvertordinary-gw5-node-1 Stopped Stderr: Container roottestbackwardcompatibilityconvertordinary-gw5-zoo3-1 Stopping Stderr: Container roottestbackwardcompatibilityconvertordinary-gw5-zoo2-1 Stopping Stderr: Container roottestbackwardcompatibilityconvertordinary-gw5-zoo1-1 Stopping Stderr: Container roottestbackwardcompatibilityconvertordinary-gw5-zoo3-1 Stopped Stderr: Container roottestbackwardcompatibilityconvertordinary-gw5-zoo2-1 Stopped Stderr: Container roottestbackwardcompatibilityconvertordinary-gw5-zoo1-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_backward_compatibility/_instances-convert_ordinary-0-gw5/node/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_backward_compatibility/_instances-convert_ordinary-0-gw5/node/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_backward_compatibility/_instances-convert_ordinary-0-gw5/.env --project-name roottestbackwardcompatibilityconvertordinary-gw5 --file /ClickHouse/tests/integration/test_backward_compatibility/_instances-convert_ordinary-0-gw5/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml down --volumes] Skipping L2SquaredNorm Checking svg('foo') Executing query select svg('foo') on backward Skipping svg Checking IPv4NumToString('foo') Executing query select IPv4NumToString('foo') on backward Skipping IPv4NumToString Checking timezoneOffset('foo') Executing query select timezoneOffset('foo') on backward Skipping timezoneOffset Checking hostName('foo') Executing query select hostName('foo') on backward Skipping hostName Checking L1Norm('foo') Executing query select L1Norm('foo') on backward Skipping L1Norm Checking tuplePlus('foo') Executing query select tuplePlus('foo') on backward Skipping tuplePlus Checking connectionid('foo') Executing query select connectionid('foo') on backward Skipping connectionid Checking modulo('foo') Executing query select modulo('foo') on backward Skipping modulo Checking simpleJSONHas('foo') Executing query select simpleJSONHas('foo') on backward Skipping simpleJSONHas Checking L2Distance('foo') Executing query select L2Distance('foo') on backward Skipping L2Distance Checking dotProduct('foo') Executing query select dotProduct('foo') on backward Stderr: Container roottestbackwardcompatibilityconvertordinary-gw5-node-1 Stopping Stderr: Container roottestbackwardcompatibilityconvertordinary-gw5-node-1 Stopped Stderr: Container roottestbackwardcompatibilityconvertordinary-gw5-node-1 Removing Stderr: Container roottestbackwardcompatibilityconvertordinary-gw5-node-1 Removed Stderr: Container roottestbackwardcompatibilityconvertordinary-gw5-zoo3-1 Stopping Stderr: Container roottestbackwardcompatibilityconvertordinary-gw5-zoo1-1 Stopping Stderr: Container roottestbackwardcompatibilityconvertordinary-gw5-zoo2-1 Stopping Stderr: Container roottestbackwardcompatibilityconvertordinary-gw5-zoo3-1 Stopped Stderr: Container roottestbackwardcompatibilityconvertordinary-gw5-zoo3-1 Removing Stderr: Container roottestbackwardcompatibilityconvertordinary-gw5-zoo2-1 Stopped Stderr: Container roottestbackwardcompatibilityconvertordinary-gw5-zoo2-1 Removing Stderr: Container roottestbackwardcompatibilityconvertordinary-gw5-zoo1-1 Stopped Stderr: Container roottestbackwardcompatibilityconvertordinary-gw5-zoo1-1 Removing Stderr: Container roottestbackwardcompatibilityconvertordinary-gw5-zoo1-1 Removed Stderr: Container roottestbackwardcompatibilityconvertordinary-gw5-zoo2-1 Removed Stderr: Container roottestbackwardcompatibilityconvertordinary-gw5-zoo3-1 Removed Stderr: Network roottestbackwardcompatibilityconvertordinary-gw5_default Removing Stderr: Network roottestbackwardcompatibilityconvertordinary-gw5_default Removed Cleanup called Docker networks for project roottestbackwardcompatibilityconvertordinary-gw5 are NETWORK ID NAME DRIVER SCOPE Skipping dotProduct Checking L2Norm('foo') Executing query select L2Norm('foo') on backward Docker containers for project roottestbackwardcompatibilityconvertordinary-gw5 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestbackwardcompatibilityconvertordinary-gw5 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestbackwardcompatibilityconvertordinary-gw5-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestbackwardcompatibilityconvertordinary-gw5 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Skipping L2Norm Checking simpleJSONExtractBool('foo') Executing query select simpleJSONExtractBool('foo') on backward Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:1 Volumes pruned: 1 test_compression_nested_columns/test.py::test_nested_compression_codec Running tests in /ClickHouse/tests/integration/test_compression_nested_columns/test.py Cluster start called. is_up=False Docker networks for project roottestcompressionnestedcolumns-gw5 are NETWORK ID NAME DRIVER SCOPE Skipping simpleJSONExtractBool Checking L2Normalize('foo') Executing query select L2Normalize('foo') on backward Docker containers for project roottestcompressionnestedcolumns-gw5 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestcompressionnestedcolumns-gw5 are DRIVER VOLUME NAME Cleanup called Docker networks for project roottestcompressionnestedcolumns-gw5 are NETWORK ID NAME DRIVER SCOPE Skipping L2Normalize Checking regexpExtract('foo') Executing query select regexpExtract('foo') on backward Docker containers for project roottestcompressionnestedcolumns-gw5 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestcompressionnestedcolumns-gw5 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestcompressionnestedcolumns-gw5-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestcompressionnestedcolumns-gw5 Trying to prune unused networks... Trying to prune unused images... Skipping regexpExtract Command:[docker image prune -f] Functions: 1172, failed: 0, skipped: 890, passed: 282 Command:[docker compose --env-file /ClickHouse/tests/integration/test_backward_compatibility/_instances-functions-0-gw9/.env --project-name roottestbackwardcompatibilityfunctions-gw9 --file /ClickHouse/tests/integration/test_backward_compatibility/_instances-functions-0-gw9/upstream/docker-compose.yml --file /ClickHouse/tests/integration/test_backward_compatibility/_instances-functions-0-gw9/backward/docker-compose.yml stop --timeout 20] [gw9] PASSED test_backward_compatibility/test_functions.py::test_string_functions Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:1 Volumes pruned: 1 Setup directory for instance: node1 Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files [] to /ClickHouse/tests/integration/test_compression_nested_columns/_instances-0-gw5/node1/configs/config.d Setup database dir /ClickHouse/tests/integration/test_compression_nested_columns/_instances-0-gw5/node1/database Setup logs dir /ClickHouse/tests/integration/test_compression_nested_columns/_instances-0-gw5/node1/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Setup directory for instance: node2 Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files [] to /ClickHouse/tests/integration/test_compression_nested_columns/_instances-0-gw5/node2/configs/config.d Setup database dir /ClickHouse/tests/integration/test_compression_nested_columns/_instances-0-gw5/node2/database Setup logs dir /ClickHouse/tests/integration/test_compression_nested_columns/_instances-0-gw5/node2/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'keeper_binary': '/clickhouse', 'keeper_cmd_prefix': 'clickhouse keeper', 'image': 'altinityinfra/integration-test:8b2301119731', 'user': '0', 'keeper_fs': 'bind', 'keeper_logs_dir1': '/ClickHouse/tests/integration/test_compression_nested_columns/_instances-0-gw5/keeper1/log', 'keeper_config_dir1': '/ClickHouse/tests/integration/test_compression_nested_columns/_instances-0-gw5/keeper1/config', 'keeper_db_dir1': '/ClickHouse/tests/integration/test_compression_nested_columns/_instances-0-gw5/keeper1/coordination', 'keeper_logs_dir2': '/ClickHouse/tests/integration/test_compression_nested_columns/_instances-0-gw5/keeper2/log', 'keeper_config_dir2': '/ClickHouse/tests/integration/test_compression_nested_columns/_instances-0-gw5/keeper2/config', 'keeper_db_dir2': '/ClickHouse/tests/integration/test_compression_nested_columns/_instances-0-gw5/keeper2/coordination', 'keeper_logs_dir3': '/ClickHouse/tests/integration/test_compression_nested_columns/_instances-0-gw5/keeper3/log', 'keeper_config_dir3': '/ClickHouse/tests/integration/test_compression_nested_columns/_instances-0-gw5/keeper3/config', 'keeper_db_dir3': '/ClickHouse/tests/integration/test_compression_nested_columns/_instances-0-gw5/keeper3/coordination'} stored in /ClickHouse/tests/integration/test_compression_nested_columns/_instances-0-gw5/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_compression_nested_columns/_instances-0-gw5/.env --project-name roottestcompressionnestedcolumns-gw5 --file /ClickHouse/tests/integration/test_compression_nested_columns/_instances-0-gw5/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_compression_nested_columns/_instances-0-gw5/node2/docker-compose.yml pull] Stderr: Container roottestbackwardcompatibilityfunctions-gw9-upstream-1 Stopping Stderr: Container roottestbackwardcompatibilityfunctions-gw9-backward-1 Stopping Stderr: Container roottestbackwardcompatibilityfunctions-gw9-backward-1 Stopped Stderr: Container roottestbackwardcompatibilityfunctions-gw9-upstream-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_backward_compatibility/_instances-functions-0-gw9/upstream/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_backward_compatibility/_instances-functions-0-gw9/upstream/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_backward_compatibility/_instances-functions-0-gw9/backward/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_backward_compatibility/_instances-functions-0-gw9/backward/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_backward_compatibility/_instances-functions-0-gw9/.env --project-name roottestbackwardcompatibilityfunctions-gw9 --file /ClickHouse/tests/integration/test_backward_compatibility/_instances-functions-0-gw9/upstream/docker-compose.yml --file /ClickHouse/tests/integration/test_backward_compatibility/_instances-functions-0-gw9/backward/docker-compose.yml down --volumes] Stderr: Container roottestbackwardcompatibilityfunctions-gw9-backward-1 Stopping Stderr: Container roottestbackwardcompatibilityfunctions-gw9-upstream-1 Stopping Stderr: Container roottestbackwardcompatibilityfunctions-gw9-backward-1 Stopped Stderr: Container roottestbackwardcompatibilityfunctions-gw9-backward-1 Removing Stderr: Container roottestbackwardcompatibilityfunctions-gw9-upstream-1 Stopped Stderr: Container roottestbackwardcompatibilityfunctions-gw9-upstream-1 Removing Stderr: Container roottestbackwardcompatibilityfunctions-gw9-backward-1 Removed Stderr: Container roottestbackwardcompatibilityfunctions-gw9-upstream-1 Removed Stderr: Network roottestbackwardcompatibilityfunctions-gw9_default Removing Stderr: Network roottestbackwardcompatibilityfunctions-gw9_default Removed Cleanup called Docker networks for project roottestbackwardcompatibilityfunctions-gw9 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestbackwardcompatibilityfunctions-gw9 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestbackwardcompatibilityfunctions-gw9 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestbackwardcompatibilityfunctions-gw9-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestbackwardcompatibilityfunctions-gw9 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:1 Volumes pruned: 1 Running tests in /ClickHouse/tests/integration/test_backward_compatibility/test_select_aggregate_alias_column.py Cluster start called. is_up=False test_backward_compatibility/test_select_aggregate_alias_column.py::test_select_aggregate_alias_column Docker networks for project roottestbackwardcompatibilityselectaggregatealiascolumn-gw9 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestbackwardcompatibilityselectaggregatealiascolumn-gw9 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestbackwardcompatibilityselectaggregatealiascolumn-gw9 are DRIVER VOLUME NAME Cleanup called Docker networks for project roottestbackwardcompatibilityselectaggregatealiascolumn-gw9 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestbackwardcompatibilityselectaggregatealiascolumn-gw9 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestbackwardcompatibilityselectaggregatealiascolumn-gw9 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestbackwardcompatibilityselectaggregatealiascolumn-gw9-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestbackwardcompatibilityselectaggregatealiascolumn-gw9 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:1 Volumes pruned: 1 Setup directory for instance: node1 Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files [] to /ClickHouse/tests/integration/test_backward_compatibility/_instances-select_aggregate_alias_column-0-gw9/node1/configs/config.d Setup database dir /ClickHouse/tests/integration/test_backward_compatibility/_instances-select_aggregate_alias_column-0-gw9/node1/database Setup logs dir /ClickHouse/tests/integration/test_backward_compatibility/_instances-select_aggregate_alias_column-0-gw9/node1/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Setup directory for instance: node2 Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files [] to /ClickHouse/tests/integration/test_backward_compatibility/_instances-select_aggregate_alias_column-0-gw9/node2/configs/config.d Setup database dir /ClickHouse/tests/integration/test_backward_compatibility/_instances-select_aggregate_alias_column-0-gw9/node2/database Setup logs dir /ClickHouse/tests/integration/test_backward_compatibility/_instances-select_aggregate_alias_column-0-gw9/node2/logs Entrypoint cmd: bash -c "trap 'pkill tail' INT TERM; clickhouse server --config-file=/etc/clickhouse-server/config.xml --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log --daemon -- ; coproc tail -f /dev/null; wait $$!" Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw'} stored in /ClickHouse/tests/integration/test_backward_compatibility/_instances-select_aggregate_alias_column-0-gw9/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_backward_compatibility/_instances-select_aggregate_alias_column-0-gw9/.env --project-name roottestbackwardcompatibilityselectaggregatealiascolumn-gw9 --file /ClickHouse/tests/integration/test_backward_compatibility/_instances-select_aggregate_alias_column-0-gw9/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_backward_compatibility/_instances-select_aggregate_alias_column-0-gw9/node2/docker-compose.yml pull] Stderr: node1 Pulling Stderr: node2 Pulling Stderr: node1 Pulled Stderr: node2 Pulled ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_backward_compatibility/_instances-select_aggregate_alias_column-0-gw9/.env --project-name roottestbackwardcompatibilityselectaggregatealiascolumn-gw9 --file /ClickHouse/tests/integration/test_backward_compatibility/_instances-select_aggregate_alias_column-0-gw9/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_backward_compatibility/_instances-select_aggregate_alias_column-0-gw9/node2/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_backward_compatibility/_instances-select_aggregate_alias_column-0-gw9/.env --project-name roottestbackwardcompatibilityselectaggregatealiascolumn-gw9 --file /ClickHouse/tests/integration/test_backward_compatibility/_instances-select_aggregate_alias_column-0-gw9/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_backward_compatibility/_instances-select_aggregate_alias_column-0-gw9/node2/docker-compose.yml up -d --no-recreate] Stderr: node2 Skipped - Image is already being pulled by zoo3 Stderr: node1 Skipped - Image is already being pulled by zoo3 Stderr: zoo1 Skipped - Image is already being pulled by zoo3 Stderr: zoo2 Skipped - Image is already being pulled by zoo3 Stderr: zoo3 Pulling Stderr: zoo3 Pulled Setup ZooKeeper Creating internal ZooKeeper dirs: ['/ClickHouse/tests/integration/test_compression_nested_columns/_instances-0-gw5/keeper1/log', '/ClickHouse/tests/integration/test_compression_nested_columns/_instances-0-gw5/keeper1/config', '/ClickHouse/tests/integration/test_compression_nested_columns/_instances-0-gw5/keeper1/coordination', '/ClickHouse/tests/integration/test_compression_nested_columns/_instances-0-gw5/keeper2/log', '/ClickHouse/tests/integration/test_compression_nested_columns/_instances-0-gw5/keeper2/config', '/ClickHouse/tests/integration/test_compression_nested_columns/_instances-0-gw5/keeper2/coordination', '/ClickHouse/tests/integration/test_compression_nested_columns/_instances-0-gw5/keeper3/log', '/ClickHouse/tests/integration/test_compression_nested_columns/_instances-0-gw5/keeper3/config', '/ClickHouse/tests/integration/test_compression_nested_columns/_instances-0-gw5/keeper3/coordination'] Command:[docker compose --project-name roottestcompressionnestedcolumns-gw5 --env-file /ClickHouse/tests/integration/test_compression_nested_columns/_instances-0-gw5/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --verbose up -d] Stderr: Network roottestbackwardcompatibilityselectaggregatealiascolumn-gw9_default Creating Stderr: Network roottestbackwardcompatibilityselectaggregatealiascolumn-gw9_default Created Stderr: Container roottestbackwardcompatibilityselectaggregatealiascolumn-gw9-node1-1 Creating Stderr: Container roottestbackwardcompatibilityselectaggregatealiascolumn-gw9-node2-1 Creating Stderr: Container roottestbackwardcompatibilityselectaggregatealiascolumn-gw9-node2-1 Created Stderr: Container roottestbackwardcompatibilityselectaggregatealiascolumn-gw9-node1-1 Created Stderr: Container roottestbackwardcompatibilityselectaggregatealiascolumn-gw9-node1-1 Starting Stderr: Container roottestbackwardcompatibilityselectaggregatealiascolumn-gw9-node2-1 Starting Stderr: Container roottestbackwardcompatibilityselectaggregatealiascolumn-gw9-node1-1 Started Stderr: Container roottestbackwardcompatibilityselectaggregatealiascolumn-gw9-node2-1 Started ClickHouse instance created get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottestbackwardcompatibilityselectaggregatealiascolumn-gw9-node1-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottestbackwardcompatibilityselectaggregatealiascolumn-gw9-node1-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node1, ip: 172.16.1.3... http://localhost:None "GET /v1.46/containers/roottestbackwardcompatibilityselectaggregatealiascolumn-gw9-node1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/f7a6014d61d0746fd21506e336ce4f6f2f007b3098744eb403719aa68ec772cb/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/f7a6014d61d0746fd21506e336ce4f6f2f007b3098744eb403719aa68ec772cb/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/f7a6014d61d0746fd21506e336ce4f6f2f007b3098744eb403719aa68ec772cb/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/f7a6014d61d0746fd21506e336ce4f6f2f007b3098744eb403719aa68ec772cb/json HTTP/1.1" 200 None Stderr:time="2025-04-02T03:36:33Z" level=trace msg="Docker Desktop integration not enabled" Stderr: Network roottestcompressionnestedcolumns-gw5_default Creating Stderr: Network roottestcompressionnestedcolumns-gw5_default Created Stderr: Container roottestcompressionnestedcolumns-gw5-zoo2-1 Creating Stderr: Container roottestcompressionnestedcolumns-gw5-zoo3-1 Creating Stderr: Container roottestcompressionnestedcolumns-gw5-zoo1-1 Creating Stderr: Container roottestcompressionnestedcolumns-gw5-zoo3-1 Created Stderr: Container roottestcompressionnestedcolumns-gw5-zoo2-1 Created Stderr: Container roottestcompressionnestedcolumns-gw5-zoo1-1 Created Stderr: Container roottestcompressionnestedcolumns-gw5-zoo2-1 Starting Stderr: Container roottestcompressionnestedcolumns-gw5-zoo3-1 Starting Stderr: Container roottestcompressionnestedcolumns-gw5-zoo1-1 Starting Stderr: Container roottestcompressionnestedcolumns-gw5-zoo1-1 Started Stderr: Container roottestcompressionnestedcolumns-gw5-zoo3-1 Started Stderr: Container roottestcompressionnestedcolumns-gw5-zoo2-1 Started Stderr:time="2025-04-02T03:36:34Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T03:36:34Z" level=debug msg="otel error" error="" Wait ZooKeeper to start get_instance_ip instance_name=zoo1 http://localhost:None "GET /v1.46/containers/roottestcompressionnestedcolumns-gw5-zoo1-1/json HTTP/1.1" 200 None get_kazoo_client: zoo1, ip:172.16.2.2, port:2181, use_ssl:False Connecting to 172.16.2.2(172.16.2.2):2181, use_ssl: False Connection dropped: socket connection error: Connection refused http://localhost:None "GET /v1.46/containers/f7a6014d61d0746fd21506e336ce4f6f2f007b3098744eb403719aa68ec772cb/json HTTP/1.1" 200 None Connecting to 172.16.2.2(172.16.2.2):2181, use_ssl: False Connection dropped: socket connection error: Connection refused ClickHouse node1 started get_instance_ip instance_name=node2 http://localhost:None "GET /v1.46/containers/roottestbackwardcompatibilityselectaggregatealiascolumn-gw9-node2-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node2 http://localhost:None "GET /v1.46/containers/roottestbackwardcompatibilityselectaggregatealiascolumn-gw9-node2-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node2, ip: 172.16.1.2... http://localhost:None "GET /v1.46/containers/roottestbackwardcompatibilityselectaggregatealiascolumn-gw9-node2-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/c43586b757d34e192ca61eb7b99effde965eae4f2b8e08f099eae711cfec9264/json HTTP/1.1" 200 None ClickHouse node2 started Executing query create table tab (x UInt64, x_alias UInt64 ALIAS x) engine = Memory on node1 Executing query create table tab (x UInt64, x_alias UInt64 ALIAS x) engine = Memory on node2 Connecting to 172.16.2.2(172.16.2.2):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Executing query insert into tab values (1) on node1 Executing query insert into tab values (1) on node2 Executing query select sum(x_alias) from remote('node{1,2}', default, tab) on node1 Connecting to 172.16.2.2(172.16.2.2):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Executing query select sum(x_alias) from remote('node{1,2}', default, tab) on node2 Executing query drop table tab on node1 Executing query drop table tab on node2 Command:[docker compose --env-file /ClickHouse/tests/integration/test_backward_compatibility/_instances-select_aggregate_alias_column-0-gw9/.env --project-name roottestbackwardcompatibilityselectaggregatealiascolumn-gw9 --file /ClickHouse/tests/integration/test_backward_compatibility/_instances-select_aggregate_alias_column-0-gw9/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_backward_compatibility/_instances-select_aggregate_alias_column-0-gw9/node2/docker-compose.yml stop --timeout 20] [gw9] PASSED test_backward_compatibility/test_select_aggregate_alias_column.py::test_select_aggregate_alias_column Connecting to 172.16.2.2(172.16.2.2):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Connection dropped: socket connection error: None Connecting to 172.16.2.2(172.16.2.2):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Connecting to 172.16.2.2(172.16.2.2):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED get_instance_ip instance_name=zoo2 http://localhost:None "GET /v1.46/containers/roottestcompressionnestedcolumns-gw5-zoo2-1/json HTTP/1.1" 200 None get_kazoo_client: zoo2, ip:172.16.2.3, port:2181, use_ssl:False Connecting to 172.16.2.3(172.16.2.3):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED get_instance_ip instance_name=zoo3 http://localhost:None "GET /v1.46/containers/roottestcompressionnestedcolumns-gw5-zoo3-1/json HTTP/1.1" 200 None get_kazoo_client: zoo3, ip:172.16.2.4, port:2181, use_ssl:False Connecting to 172.16.2.4(172.16.2.4):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED All instances of ZooKeeper started: ('zoo1', 'zoo2', 'zoo3') ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_compression_nested_columns/_instances-0-gw5/.env --project-name roottestcompressionnestedcolumns-gw5 --file /ClickHouse/tests/integration/test_compression_nested_columns/_instances-0-gw5/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_compression_nested_columns/_instances-0-gw5/node2/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_compression_nested_columns/_instances-0-gw5/.env --project-name roottestcompressionnestedcolumns-gw5 --file /ClickHouse/tests/integration/test_compression_nested_columns/_instances-0-gw5/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_compression_nested_columns/_instances-0-gw5/node2/docker-compose.yml up -d --no-recreate] Stderr: Container roottestcompressionnestedcolumns-gw5-zoo3-1 Running Stderr: Container roottestcompressionnestedcolumns-gw5-zoo2-1 Running Stderr: Container roottestcompressionnestedcolumns-gw5-zoo1-1 Running Stderr: Container roottestcompressionnestedcolumns-gw5-node2-1 Creating Stderr: Container roottestcompressionnestedcolumns-gw5-node1-1 Creating Stderr: Container roottestcompressionnestedcolumns-gw5-node2-1 Created Stderr: Container roottestcompressionnestedcolumns-gw5-node1-1 Created Stderr: Container roottestcompressionnestedcolumns-gw5-node1-1 Starting Stderr: Container roottestcompressionnestedcolumns-gw5-node2-1 Starting Stderr: Container roottestcompressionnestedcolumns-gw5-node2-1 Started Stderr: Container roottestcompressionnestedcolumns-gw5-node1-1 Started ClickHouse instance created get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottestcompressionnestedcolumns-gw5-node1-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottestcompressionnestedcolumns-gw5-node1-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node1, ip: 172.16.2.6... http://localhost:None "GET /v1.46/containers/roottestcompressionnestedcolumns-gw5-node1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/24017f26a7097b2e7df18c91eeeac16b3c78ab4d60d3fdb7ba3f10859c032b27/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/24017f26a7097b2e7df18c91eeeac16b3c78ab4d60d3fdb7ba3f10859c032b27/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/24017f26a7097b2e7df18c91eeeac16b3c78ab4d60d3fdb7ba3f10859c032b27/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/24017f26a7097b2e7df18c91eeeac16b3c78ab4d60d3fdb7ba3f10859c032b27/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/24017f26a7097b2e7df18c91eeeac16b3c78ab4d60d3fdb7ba3f10859c032b27/json HTTP/1.1" 200 None ClickHouse node1 started get_instance_ip instance_name=node2 http://localhost:None "GET /v1.46/containers/roottestcompressionnestedcolumns-gw5-node2-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node2 http://localhost:None "GET /v1.46/containers/roottestcompressionnestedcolumns-gw5-node2-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node2, ip: 172.16.2.5... http://localhost:None "GET /v1.46/containers/roottestcompressionnestedcolumns-gw5-node2-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b90324aab393d4d5c3719f17ed70a3fe304b94e88c85545a4a6888fcd12d6215/json HTTP/1.1" 200 None ClickHouse node2 started Executing query CREATE TABLE compression_table ( key UInt64, column_ok Nullable(UInt64) CODEC(Delta, LZ4), column_array Array(Array(UInt64)) CODEC(T64, LZ4), column_bad LowCardinality(Int64) CODEC(Delta) ) ENGINE = ReplicatedMergeTree('/t', '0') ORDER BY tuple() PARTITION BY key SETTINGS min_rows_for_wide_part = 0, min_bytes_for_wide_part = 0, replace_long_file_name_to_hash = 0; on node1 Stderr: Container roottestbackwardcompatibilityselectaggregatealiascolumn-gw9-node1-1 Stopping Stderr: Container roottestbackwardcompatibilityselectaggregatealiascolumn-gw9-node2-1 Stopping Stderr: Container roottestbackwardcompatibilityselectaggregatealiascolumn-gw9-node2-1 Stopped Stderr: Container roottestbackwardcompatibilityselectaggregatealiascolumn-gw9-node1-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_backward_compatibility/_instances-select_aggregate_alias_column-0-gw9/node1/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_backward_compatibility/_instances-select_aggregate_alias_column-0-gw9/node1/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_backward_compatibility/_instances-select_aggregate_alias_column-0-gw9/node2/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_backward_compatibility/_instances-select_aggregate_alias_column-0-gw9/node2/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_backward_compatibility/_instances-select_aggregate_alias_column-0-gw9/.env --project-name roottestbackwardcompatibilityselectaggregatealiascolumn-gw9 --file /ClickHouse/tests/integration/test_backward_compatibility/_instances-select_aggregate_alias_column-0-gw9/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_backward_compatibility/_instances-select_aggregate_alias_column-0-gw9/node2/docker-compose.yml down --volumes] Executing query CREATE TABLE compression_table ( key UInt64, column_ok Nullable(UInt64) CODEC(Delta, LZ4), column_array Array(Array(UInt64)) CODEC(T64, LZ4), column_bad LowCardinality(Int64) CODEC(Delta) ) ENGINE = ReplicatedMergeTree('/t', '1') ORDER BY tuple() PARTITION BY key SETTINGS min_rows_for_wide_part = 0, min_bytes_for_wide_part = 0, replace_long_file_name_to_hash = 0; on node2 Executing query INSERT INTO compression_table VALUES (1, 1, [[77]], 32) on node1 Executing query SYSTEM SYNC REPLICA compression_table on node2 Executing query DETACH TABLE compression_table on node1 Executing query DETACH TABLE compression_table on node2 Executing query ATTACH TABLE compression_table on node1 Stderr: Container roottestbackwardcompatibilityselectaggregatealiascolumn-gw9-node1-1 Stopping Stderr: Container roottestbackwardcompatibilityselectaggregatealiascolumn-gw9-node2-1 Stopping Stderr: Container roottestbackwardcompatibilityselectaggregatealiascolumn-gw9-node1-1 Stopped Stderr: Container roottestbackwardcompatibilityselectaggregatealiascolumn-gw9-node1-1 Removing Stderr: Container roottestbackwardcompatibilityselectaggregatealiascolumn-gw9-node2-1 Stopped Stderr: Container roottestbackwardcompatibilityselectaggregatealiascolumn-gw9-node2-1 Removing Stderr: Container roottestbackwardcompatibilityselectaggregatealiascolumn-gw9-node1-1 Removed Stderr: Container roottestbackwardcompatibilityselectaggregatealiascolumn-gw9-node2-1 Removed Stderr: Network roottestbackwardcompatibilityselectaggregatealiascolumn-gw9_default Removing Stderr: Network roottestbackwardcompatibilityselectaggregatealiascolumn-gw9_default Removed Cleanup called Executing query ATTACH TABLE compression_table on node2 Docker networks for project roottestbackwardcompatibilityselectaggregatealiascolumn-gw9 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestbackwardcompatibilityselectaggregatealiascolumn-gw9 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestbackwardcompatibilityselectaggregatealiascolumn-gw9 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestbackwardcompatibilityselectaggregatealiascolumn-gw9-.*-1$' --format '{{.ID}}:{{.Names}}'] run container_id:roottestcompressionnestedcolumns-gw5-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "tail -c +17 /var/lib/clickhouse/data/default/compression_table/1_0_0_0/column_ok.bin | od -x -N 1 | head -n 1 | awk '{print $2}'"] Command:[docker exec roottestcompressionnestedcolumns-gw5-node1-1 bash -c tail -c +17 /var/lib/clickhouse/data/default/compression_table/1_0_0_0/column_ok.bin | od -x -N 1 | head -n 1 | awk '{print $2}'] Unstopped containers: {} No running containers for project: roottestbackwardcompatibilityselectaggregatealiascolumn-gw9 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:0091 run container_id:roottestcompressionnestedcolumns-gw5-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "tail -c +17 /var/lib/clickhouse/data/default/compression_table/1_0_0_0/column_ok.null.bin | od -x -N 1 | head -n 1 | awk '{print $2}'"] Command:[docker exec roottestcompressionnestedcolumns-gw5-node1-1 bash -c tail -c +17 /var/lib/clickhouse/data/default/compression_table/1_0_0_0/column_ok.null.bin | od -x -N 1 | head -n 1 | awk '{print $2}'] Stdout:1 Volumes pruned: 1 Stdout:0082 run container_id:roottestcompressionnestedcolumns-gw5-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "tail -c +17 /var/lib/clickhouse/data/default/compression_table/1_0_0_0/column_array.bin | od -x -N 1 | head -n 1 | awk '{print $2}'"] Command:[docker exec roottestcompressionnestedcolumns-gw5-node1-1 bash -c tail -c +17 /var/lib/clickhouse/data/default/compression_table/1_0_0_0/column_array.bin | od -x -N 1 | head -n 1 | awk '{print $2}'] Stdout:0091 run container_id:roottestcompressionnestedcolumns-gw5-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "tail -c +17 /var/lib/clickhouse/data/default/compression_table/1_0_0_0/column_array.size0.bin | od -x -N 1 | head -n 1 | awk '{print $2}'"] Command:[docker exec roottestcompressionnestedcolumns-gw5-node2-1 bash -c tail -c +17 /var/lib/clickhouse/data/default/compression_table/1_0_0_0/column_array.size0.bin | od -x -N 1 | head -n 1 | awk '{print $2}'] Stdout:0082 run container_id:roottestcompressionnestedcolumns-gw5-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "tail -c +17 /var/lib/clickhouse/data/default/compression_table/1_0_0_0/column_array.size1.bin | od -x -N 1 | head -n 1 | awk '{print $2}'"] Command:[docker exec roottestcompressionnestedcolumns-gw5-node2-1 bash -c tail -c +17 /var/lib/clickhouse/data/default/compression_table/1_0_0_0/column_array.size1.bin | od -x -N 1 | head -n 1 | awk '{print $2}'] Stdout:0082 run container_id:roottestcompressionnestedcolumns-gw5-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "tail -c +17 /var/lib/clickhouse/data/default/compression_table/1_0_0_0/column_bad.dict.bin | od -x -N 1 | head -n 1 | awk '{print $2}'"] Command:[docker exec roottestcompressionnestedcolumns-gw5-node2-1 bash -c tail -c +17 /var/lib/clickhouse/data/default/compression_table/1_0_0_0/column_bad.dict.bin | od -x -N 1 | head -n 1 | awk '{print $2}'] Stdout:0092 run container_id:roottestcompressionnestedcolumns-gw5-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "tail -c +17 /var/lib/clickhouse/data/default/compression_table/1_0_0_0/column_bad.bin | od -x -N 1 | head -n 1 | awk '{print $2}'"] Command:[docker exec roottestcompressionnestedcolumns-gw5-node1-1 bash -c tail -c +17 /var/lib/clickhouse/data/default/compression_table/1_0_0_0/column_bad.bin | od -x -N 1 | head -n 1 | awk '{print $2}'] Stdout:0002 run container_id:roottestcompressionnestedcolumns-gw5-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "tail -c +17 /var/lib/clickhouse/data/default/compression_table/1_0_0_0/column_ok.bin | od -x -N 1 | head -n 1 | awk '{print $2}'"] Command:[docker exec roottestcompressionnestedcolumns-gw5-node2-1 bash -c tail -c +17 /var/lib/clickhouse/data/default/compression_table/1_0_0_0/column_ok.bin | od -x -N 1 | head -n 1 | awk '{print $2}'] Stdout:0091 run container_id:roottestcompressionnestedcolumns-gw5-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "tail -c +17 /var/lib/clickhouse/data/default/compression_table/1_0_0_0/column_ok.null.bin | od -x -N 1 | head -n 1 | awk '{print $2}'"] Command:[docker exec roottestcompressionnestedcolumns-gw5-node2-1 bash -c tail -c +17 /var/lib/clickhouse/data/default/compression_table/1_0_0_0/column_ok.null.bin | od -x -N 1 | head -n 1 | awk '{print $2}'] Stdout:0082 run container_id:roottestcompressionnestedcolumns-gw5-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "tail -c +17 /var/lib/clickhouse/data/default/compression_table/1_0_0_0/column_array.bin | od -x -N 1 | head -n 1 | awk '{print $2}'"] Command:[docker exec roottestcompressionnestedcolumns-gw5-node1-1 bash -c tail -c +17 /var/lib/clickhouse/data/default/compression_table/1_0_0_0/column_array.bin | od -x -N 1 | head -n 1 | awk '{print $2}'] Stdout:0091 run container_id:roottestcompressionnestedcolumns-gw5-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "tail -c +17 /var/lib/clickhouse/data/default/compression_table/1_0_0_0/column_array.size0.bin | od -x -N 1 | head -n 1 | awk '{print $2}'"] Command:[docker exec roottestcompressionnestedcolumns-gw5-node2-1 bash -c tail -c +17 /var/lib/clickhouse/data/default/compression_table/1_0_0_0/column_array.size0.bin | od -x -N 1 | head -n 1 | awk '{print $2}'] Stdout:0082 run container_id:roottestcompressionnestedcolumns-gw5-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "tail -c +17 /var/lib/clickhouse/data/default/compression_table/1_0_0_0/column_array.size1.bin | od -x -N 1 | head -n 1 | awk '{print $2}'"] Command:[docker exec roottestcompressionnestedcolumns-gw5-node2-1 bash -c tail -c +17 /var/lib/clickhouse/data/default/compression_table/1_0_0_0/column_array.size1.bin | od -x -N 1 | head -n 1 | awk '{print $2}'] Stdout:0082 run container_id:roottestcompressionnestedcolumns-gw5-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "tail -c +17 /var/lib/clickhouse/data/default/compression_table/1_0_0_0/column_bad.dict.bin | od -x -N 1 | head -n 1 | awk '{print $2}'"] Command:[docker exec roottestcompressionnestedcolumns-gw5-node2-1 bash -c tail -c +17 /var/lib/clickhouse/data/default/compression_table/1_0_0_0/column_bad.dict.bin | od -x -N 1 | head -n 1 | awk '{print $2}'] Stdout:0092 run container_id:roottestcompressionnestedcolumns-gw5-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "tail -c +17 /var/lib/clickhouse/data/default/compression_table/1_0_0_0/column_bad.bin | od -x -N 1 | head -n 1 | awk '{print $2}'"] Command:[docker exec roottestcompressionnestedcolumns-gw5-node1-1 bash -c tail -c +17 /var/lib/clickhouse/data/default/compression_table/1_0_0_0/column_bad.bin | od -x -N 1 | head -n 1 | awk '{print $2}'] Stdout:0002 Command:[docker compose --env-file /ClickHouse/tests/integration/test_compression_nested_columns/_instances-0-gw5/.env --project-name roottestcompressionnestedcolumns-gw5 --file /ClickHouse/tests/integration/test_compression_nested_columns/_instances-0-gw5/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_compression_nested_columns/_instances-0-gw5/node2/docker-compose.yml stop --timeout 20] [gw5] PASSED test_compression_nested_columns/test.py::test_nested_compression_codec Connection dropped: socket connection error: None Stderr: Container roottestcompressionnestedcolumns-gw5-node1-1 Stopping Stderr: Container roottestcompressionnestedcolumns-gw5-node2-1 Stopping Stderr: Container roottestcompressionnestedcolumns-gw5-node2-1 Stopped Stderr: Container roottestcompressionnestedcolumns-gw5-node1-1 Stopped Stderr: Container roottestcompressionnestedcolumns-gw5-zoo2-1 Stopping Stderr: Container roottestcompressionnestedcolumns-gw5-zoo1-1 Stopping Stderr: Container roottestcompressionnestedcolumns-gw5-zoo3-1 Stopping Stderr: Container roottestcompressionnestedcolumns-gw5-zoo1-1 Stopped Stderr: Container roottestcompressionnestedcolumns-gw5-zoo2-1 Stopped Stderr: Container roottestcompressionnestedcolumns-gw5-zoo3-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_compression_nested_columns/_instances-0-gw5/node1/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_compression_nested_columns/_instances-0-gw5/node1/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_compression_nested_columns/_instances-0-gw5/node2/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_compression_nested_columns/_instances-0-gw5/node2/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_compression_nested_columns/_instances-0-gw5/.env --project-name roottestcompressionnestedcolumns-gw5 --file /ClickHouse/tests/integration/test_compression_nested_columns/_instances-0-gw5/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_compression_nested_columns/_instances-0-gw5/node2/docker-compose.yml down --volumes] Stderr: Container roottestcompressionnestedcolumns-gw5-node2-1 Stopping Stderr: Container roottestcompressionnestedcolumns-gw5-node1-1 Stopping Stderr: Container roottestcompressionnestedcolumns-gw5-node2-1 Stopped Stderr: Container roottestcompressionnestedcolumns-gw5-node2-1 Removing Stderr: Container roottestcompressionnestedcolumns-gw5-node1-1 Stopped Stderr: Container roottestcompressionnestedcolumns-gw5-node1-1 Removing Stderr: Container roottestcompressionnestedcolumns-gw5-node2-1 Removed Stderr: Container roottestcompressionnestedcolumns-gw5-node1-1 Removed Stderr: Container roottestcompressionnestedcolumns-gw5-zoo2-1 Stopping Stderr: Container roottestcompressionnestedcolumns-gw5-zoo3-1 Stopping Stderr: Container roottestcompressionnestedcolumns-gw5-zoo1-1 Stopping Stderr: Container roottestcompressionnestedcolumns-gw5-zoo2-1 Stopped Stderr: Container roottestcompressionnestedcolumns-gw5-zoo2-1 Removing Stderr: Container roottestcompressionnestedcolumns-gw5-zoo3-1 Stopped Stderr: Container roottestcompressionnestedcolumns-gw5-zoo3-1 Removing Stderr: Container roottestcompressionnestedcolumns-gw5-zoo1-1 Stopped Stderr: Container roottestcompressionnestedcolumns-gw5-zoo1-1 Removing Stderr: Container roottestcompressionnestedcolumns-gw5-zoo3-1 Removed Stderr: Container roottestcompressionnestedcolumns-gw5-zoo1-1 Removed Stderr: Container roottestcompressionnestedcolumns-gw5-zoo2-1 Removed Stderr: Network roottestcompressionnestedcolumns-gw5_default Removing Stderr: Network roottestcompressionnestedcolumns-gw5_default Removed Cleanup called Docker networks for project roottestcompressionnestedcolumns-gw5 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestcompressionnestedcolumns-gw5 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestcompressionnestedcolumns-gw5 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestcompressionnestedcolumns-gw5-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestcompressionnestedcolumns-gw5 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:1 Volumes pruned: 1 test_custom_settings/test.py::test_custom_settings Running tests in /ClickHouse/tests/integration/test_custom_settings/test.py Cluster start called. is_up=False Docker networks for project roottestcustomsettings-gw5 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestcustomsettings-gw5 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestcustomsettings-gw5 are DRIVER VOLUME NAME Cleanup called Docker networks for project roottestcustomsettings-gw5 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestcustomsettings-gw5 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestcustomsettings-gw5 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestcustomsettings-gw5-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestcustomsettings-gw5 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:1 Volumes pruned: 1 Setup directory for instance: node Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files [] to /ClickHouse/tests/integration/test_custom_settings/_instances-0-gw5/node/configs/config.d Setup database dir /ClickHouse/tests/integration/test_custom_settings/_instances-0-gw5/node/database Setup logs dir /ClickHouse/tests/integration/test_custom_settings/_instances-0-gw5/node/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw'} stored in /ClickHouse/tests/integration/test_custom_settings/_instances-0-gw5/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_custom_settings/_instances-0-gw5/.env --project-name roottestcustomsettings-gw5 --file /ClickHouse/tests/integration/test_custom_settings/_instances-0-gw5/node/docker-compose.yml pull] Stderr: node Pulling Stderr: node Pulled ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_custom_settings/_instances-0-gw5/.env --project-name roottestcustomsettings-gw5 --file /ClickHouse/tests/integration/test_custom_settings/_instances-0-gw5/node/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_custom_settings/_instances-0-gw5/.env --project-name roottestcustomsettings-gw5 --file /ClickHouse/tests/integration/test_custom_settings/_instances-0-gw5/node/docker-compose.yml up -d --no-recreate] Stderr: Network roottestcustomsettings-gw5_default Creating Stderr: Network roottestcustomsettings-gw5_default Created Stderr: Container roottestcustomsettings-gw5-node-1 Creating Stderr: Container roottestcustomsettings-gw5-node-1 Created Stderr: Container roottestcustomsettings-gw5-node-1 Starting Stderr: Container roottestcustomsettings-gw5-node-1 Started ClickHouse instance created get_instance_ip instance_name=node http://localhost:None "GET /v1.46/containers/roottestcustomsettings-gw5-node-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node http://localhost:None "GET /v1.46/containers/roottestcustomsettings-gw5-node-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node, ip: 172.16.1.2... http://localhost:None "GET /v1.46/containers/roottestcustomsettings-gw5-node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/70132f88b7d0845441c5c3568cc97ca5b0a560dfcda83fc051f2b608bbba9a90/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/70132f88b7d0845441c5c3568cc97ca5b0a560dfcda83fc051f2b608bbba9a90/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/70132f88b7d0845441c5c3568cc97ca5b0a560dfcda83fc051f2b608bbba9a90/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/70132f88b7d0845441c5c3568cc97ca5b0a560dfcda83fc051f2b608bbba9a90/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/70132f88b7d0845441c5c3568cc97ca5b0a560dfcda83fc051f2b608bbba9a90/json HTTP/1.1" 200 None ClickHouse node started run container_id:roottestcustomsettings-gw5-node-1 detach:False nothrow:False cmd: ['bash', '-c', 'mkdir -p $(dirname /etc/clickhouse-server/users.d/z.xml) && echo PGNsaWNraG91c2U+CiAgICA8cHJvZmlsZXM+CiAgICAgICAgPGRlZmF1bHQ+CiAgICAgICAgICAgIDxjdXN0b21fYT5JbnQ2NF8tNTwvY3VzdG9tX2E+CiAgICAgICAgICAgIDxjdXN0b21fYj5VSW50NjRfMTAwMDAwMDAwMDA8L2N1c3RvbV9iPgogICAgICAgICAgICA8Y3VzdG9tX2M+RmxvYXQ2NF8tNDMuMjVlLTE8L2N1c3RvbV9jPgogICAgICAgICAgICA8Y3VzdG9tX2Q+J3NvbWUgdGV4dCc8L2N1c3RvbV9kPgogICAgICAgIDwvZGVmYXVsdD4KICAgIDwvcHJvZmlsZXM+CjwvY2xpY2tob3VzZT4K | base64 --decode > /etc/clickhouse-server/users.d/z.xml'] Command:[docker exec roottestcustomsettings-gw5-node-1 bash -c mkdir -p $(dirname /etc/clickhouse-server/users.d/z.xml) && echo PGNsaWNraG91c2U+CiAgICA8cHJvZmlsZXM+CiAgICAgICAgPGRlZmF1bHQ+CiAgICAgICAgICAgIDxjdXN0b21fYT5JbnQ2NF8tNTwvY3VzdG9tX2E+CiAgICAgICAgICAgIDxjdXN0b21fYj5VSW50NjRfMTAwMDAwMDAwMDA8L2N1c3RvbV9iPgogICAgICAgICAgICA8Y3VzdG9tX2M+RmxvYXQ2NF8tNDMuMjVlLTE8L2N1c3RvbV9jPgogICAgICAgICAgICA8Y3VzdG9tX2Q+J3NvbWUgdGV4dCc8L2N1c3RvbV9kPgogICAgICAgIDwvZGVmYXVsdD4KICAgIDwvcHJvZmlsZXM+CjwvY2xpY2tob3VzZT4K | base64 --decode > /etc/clickhouse-server/users.d/z.xml] Executing query SYSTEM RELOAD CONFIG on node Executing query SELECT getSetting('custom_a') on node Executing query SELECT getSetting('custom_b') on node Executing query SELECT getSetting('custom_c') on node Executing query SELECT getSetting('custom_d') on node Command:[docker compose --env-file /ClickHouse/tests/integration/test_custom_settings/_instances-0-gw5/.env --project-name roottestcustomsettings-gw5 --file /ClickHouse/tests/integration/test_custom_settings/_instances-0-gw5/node/docker-compose.yml stop --timeout 20] [gw5] PASSED test_custom_settings/test.py::test_custom_settings Stderr: Container roottestcustomsettings-gw5-node-1 Stopping Stderr: Container roottestcustomsettings-gw5-node-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_custom_settings/_instances-0-gw5/node/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_custom_settings/_instances-0-gw5/node/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_custom_settings/_instances-0-gw5/.env --project-name roottestcustomsettings-gw5 --file /ClickHouse/tests/integration/test_custom_settings/_instances-0-gw5/node/docker-compose.yml down --volumes] Stderr: Container roottestcustomsettings-gw5-node-1 Stopping Stderr: Container roottestcustomsettings-gw5-node-1 Stopped Stderr: Container roottestcustomsettings-gw5-node-1 Removing Stderr: Container roottestcustomsettings-gw5-node-1 Removed Stderr: Network roottestcustomsettings-gw5_default Removing Stderr: Network roottestcustomsettings-gw5_default Removed Cleanup called Docker networks for project roottestcustomsettings-gw5 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestcustomsettings-gw5 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestcustomsettings-gw5 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestcustomsettings-gw5-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestcustomsettings-gw5 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:1 Volumes pruned: 1 Connection dropped: socket connection error: None Connection dropped: socket connection error: None Connection dropped: socket connection error: None Failed connecting to Zookeeper within the connection retry policy. Connection dropped: socket connection error: None Executing query SYSTEM SYNC REPLICA replicated_mt_1 on node1 Executing query CHECK TABLE replicated_mt_1 PARTITION 201901 on node1 Executing query SELECT count() from replicated_mt_1 on node1 Executing query SELECT path FROM system.parts WHERE table = 'replicated_mt_1' and name = '201901_0_0_0' on node2 run container_id:roottestchecktable-gw4-node2-1 detach:False nothrow:False cmd: ['bash', '-c', 'rm -r /var/lib/clickhouse/store/4ad/4adc6135-f5f8-480b-a66a-7cea8a2bf56c/201901_0_0_0//*'] Command:[docker exec --privileged roottestchecktable-gw4-node2-1 bash -c rm -r /var/lib/clickhouse/store/4ad/4adc6135-f5f8-480b-a66a-7cea8a2bf56c/201901_0_0_0//*] Executing query CHECK TABLE replicated_mt_1 PARTITION 201901 on node2 Executing query SYSTEM SYNC REPLICA replicated_mt_1 on node1 Executing query CHECK TABLE replicated_mt_1 PARTITION 201901 on node1 Executing query SELECT count() from replicated_mt_1 on node1 [gw4] PASSED test_check_table/test.py::test_check_replicated_table_corruption[-_0-.bin] test_check_table/test.py::test_check_replicated_table_simple[-_0] Executing query DROP TABLE IF EXISTS replicated_mt SYNC on node1 Executing query CREATE TABLE replicated_mt(date Date, id UInt32, value Int32) ENGINE = ReplicatedMergeTree('/clickhouse/tables/replicated_mt__0', 'node1') PARTITION BY toYYYYMM(date) ORDER BY id on node1 Executing query DROP TABLE IF EXISTS replicated_mt SYNC on node2 Executing query CREATE TABLE replicated_mt(date Date, id UInt32, value Int32) ENGINE = ReplicatedMergeTree('/clickhouse/tables/replicated_mt__0', 'node2') PARTITION BY toYYYYMM(date) ORDER BY id on node2 Executing query INSERT INTO replicated_mt VALUES (toDate('2019-02-01'), 1, 10), (toDate('2019-02-01'), 2, 12) on node1 Executing query SYSTEM SYNC REPLICA replicated_mt on node2 Executing query SELECT count() from replicated_mt on node1 Executing query SELECT count() from replicated_mt on node2 Executing query CHECK TABLE replicated_mt on node1 Executing query CHECK TABLE replicated_mt on node2 Executing query INSERT INTO replicated_mt VALUES (toDate('2019-01-02'), 3, 10), (toDate('2019-01-02'), 4, 12) on node2 Executing query SYSTEM SYNC REPLICA replicated_mt on node1 Executing query SELECT count() from replicated_mt on node1 Executing query SELECT count() from replicated_mt on node2 Executing query CHECK TABLE replicated_mt PARTITION 201901 on node1 Executing query CHECK TABLE replicated_mt PARTITION 201901 on node2 Executing query CHECK TABLE replicated_mt on node2 Executing query CHECK TABLE replicated_mt PART '201801_0_0_0' on node2 Executing query CHECK TABLE replicated_mt PART '201902_0_0_0' on node2 Command:[docker compose --env-file /ClickHouse/tests/integration/test_check_table/_instances-0-gw4/.env --project-name roottestchecktable-gw4 --file /ClickHouse/tests/integration/test_check_table/_instances-0-gw4/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_check_table/_instances-0-gw4/node2/docker-compose.yml stop --timeout 20] [gw4] PASSED test_check_table/test.py::test_check_replicated_table_simple[-_0] Stderr: Container roottestchecktable-gw4-node1-1 Stopping Stderr: Container roottestchecktable-gw4-node2-1 Stopping Stderr: Container roottestchecktable-gw4-node1-1 Stopped Stderr: Container roottestchecktable-gw4-node2-1 Stopped Stderr: Container roottestchecktable-gw4-zoo1-1 Stopping Stderr: Container roottestchecktable-gw4-zoo2-1 Stopping Stderr: Container roottestchecktable-gw4-zoo3-1 Stopping Stderr: Container roottestchecktable-gw4-zoo3-1 Stopped Stderr: Container roottestchecktable-gw4-zoo2-1 Stopped Stderr: Container roottestchecktable-gw4-zoo1-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_check_table/_instances-0-gw4/node1/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_check_table/_instances-0-gw4/node1/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_check_table/_instances-0-gw4/node2/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_check_table/_instances-0-gw4/node2/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_check_table/_instances-0-gw4/.env --project-name roottestchecktable-gw4 --file /ClickHouse/tests/integration/test_check_table/_instances-0-gw4/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_check_table/_instances-0-gw4/node2/docker-compose.yml down --volumes] Stderr: Container roottestchecktable-gw4-node1-1 Stopping Stderr: Container roottestchecktable-gw4-node2-1 Stopping Stderr: Container roottestchecktable-gw4-node1-1 Stopped Stderr: Container roottestchecktable-gw4-node1-1 Removing Stderr: Container roottestchecktable-gw4-node2-1 Stopped Stderr: Container roottestchecktable-gw4-node2-1 Removing Stderr: Container roottestchecktable-gw4-node1-1 Removed Stderr: Container roottestchecktable-gw4-node2-1 Removed Stderr: Container roottestchecktable-gw4-zoo1-1 Stopping Stderr: Container roottestchecktable-gw4-zoo3-1 Stopping Stderr: Container roottestchecktable-gw4-zoo2-1 Stopping Stderr: Container roottestchecktable-gw4-zoo3-1 Stopped Stderr: Container roottestchecktable-gw4-zoo3-1 Removing Stderr: Container roottestchecktable-gw4-zoo1-1 Stopped Stderr: Container roottestchecktable-gw4-zoo1-1 Removing Stderr: Container roottestchecktable-gw4-zoo2-1 Stopped Stderr: Container roottestchecktable-gw4-zoo2-1 Removing Stderr: Container roottestchecktable-gw4-zoo1-1 Removed Stderr: Container roottestchecktable-gw4-zoo2-1 Removed Stderr: Container roottestchecktable-gw4-zoo3-1 Removed Stderr: Network roottestchecktable-gw4_default Removing Stderr: Network roottestchecktable-gw4_default Removed Cleanup called Docker networks for project roottestchecktable-gw4 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestchecktable-gw4 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestchecktable-gw4 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestchecktable-gw4-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestchecktable-gw4 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:1 Volumes pruned: 1 test_azure_blob_storage_zero_copy_replication/test.py::test_zero_copy_replication ENV DOCKER_KERBEROS_KDC_TAG 9391ecdee8d7 ENV CLICKHOUSE_TESTS_SERVER_BIN_PATH /clickhouse ENV MSAN_OPTIONS abort_on_error=1 poison_in_dtor=1 ENV JAVA_TOOL_OPTIONS -Djdk.attach.allowAttachSelf=true ENV TSAN_OPTIONS halt_on_error=1 abort_on_error=1 history_size=7 memory_limit_mb=46080 second_deadlock_stack=1 ENV HOSTNAME cec30002cd47 ENV SHLVL 0 ENV HOME /root ENV OLDPWD / ENV DOCKER_HELPER_TAG 5dc43a6382f0 ENV PYTHONUNBUFFERED 1 ENV DOCKER_PYTHON_BOTTLE_TAG caad4729259e ENV UBSAN_OPTIONS print_stacktrace=1 ENV PYTEST_ADDOPTS --dist=loadfile -n 10 -rfEps --run-id=0 --color=no --durations=0 test_access_for_functions/test.py::test_access_rights_for_function test_access_for_functions/test.py::test_ignore_obsolete_grant_on_database test_aggregation_memory_efficient/test.py::test_remote test_allowed_client_hosts/test.py::test_allowed_host test_allowed_client_hosts/test.py::test_denied_host test_alter_codec/test.py::test_alter_codec_index test_alter_codec/test.py::test_alter_codec_pk test_alter_on_mixed_type_cluster/test.py::test_alter_on_cluter_non_replicated test_alter_on_mixed_type_cluster/test.py::test_alter_replicated_on_cluster test_alternative_keeper_config/test.py::test_create_insert test_async_connect_to_multiple_ips/test.py::test test_async_metrics_in_cgroup/test.py::test_normalized_user_cpu test_async_metrics_in_cgroup/test.py::test_system_wide_metrics test_async_metrics_in_cgroup/test.py::test_user_cpu_accounting 'test_attach_partition_with_large_destination/test.py::test_attach_partition_with_large_destination[Atomic]' 'test_attach_partition_with_large_destination/test.py::test_attach_partition_with_large_destination[Ordinary]' test_attach_table_normalizer/test.py::test_attach_substr test_attach_table_normalizer/test.py::test_attach_substr_restart test_authentication/test.py::test_authentication_fail test_authentication/test.py::test_authentication_pass test_azure_blob_storage_zero_copy_replication/test.py::test_zero_copy_replication test_backup_restore/test.py::test_attach_partition test_backup_restore/test.py::test_replace_partition test_backup_restore/test.py::test_restore test_backup_restore_new/test_shutdown_wait_backup.py::test_shutdown_wait_backup 'test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-False]' 'test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-True]' 'test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[native-True]' test_backup_restore_on_cluster/test.py::test_backup_restore_on_single_replica test_backup_restore_on_cluster/test.py::test_different_tables_on_nodes test_backup_restore_on_cluster/test.py::test_empty_replicated_table test_backup_restore_on_cluster/test.py::test_file_deduplication test_backup_restore_on_cluster/test.py::test_get_error_from_other_host test_backup_restore_on_cluster/test.py::test_keeper_value_max_size test_backup_restore_on_cluster/test.py::test_mutation test_backup_restore_on_cluster/test.py::test_projection test_backup_restore_on_cluster/test.py::test_replicated_database test_backup_restore_on_cluster/test.py::test_replicated_database_async test_backup_restore_on_cluster/test.py::test_replicated_database_compare_parts 'test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[database]' 'test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[uuid]' test_backup_restore_on_cluster/test.py::test_replicated_table test_backup_restore_on_cluster/test.py::test_replicated_table_after_alters test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_bigger_cluster test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_smaller_cluster test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_def test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_insert test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_merge test_backup_restore_on_cluster/test.py::test_replicated_table_with_uuid_in_zkpath test_backup_restore_on_cluster/test.py::test_required_privileges test_backup_restore_on_cluster/test.py::test_shutdown_waits_for_backup test_backup_restore_on_cluster/test.py::test_system_functions test_backup_restore_on_cluster/test.py::test_system_users test_backup_restore_on_cluster/test.py::test_table_in_replicated_database_with_not_synced_def test_backup_restore_on_cluster/test.py::test_table_with_parts_in_queue_considered_non_empty test_backup_restore_on_cluster/test.py::test_tables_dependency test_backward_compatibility/test_convert_ordinary.py::test_convert_ordinary_to_atomic test_backward_compatibility/test_cte_distributed.py::test_cte_distributed test_backward_compatibility/test_functions.py::test_aggregate_states test_backward_compatibility/test_functions.py::test_string_functions test_backward_compatibility/test_insert_profile_events.py::test_new_client_compatible test_backward_compatibility/test_insert_profile_events.py::test_old_client_compatible test_backward_compatibility/test_normalized_count_comparison.py::test_select_aggregate_alias_column test_backward_compatibility/test_select_aggregate_alias_column.py::test_select_aggregate_alias_column test_backward_compatibility/test_vertical_merges_from_compact_parts.py::test_vertical_merges_from_compact_parts test_broken_part_during_merge/test.py::test_merge_and_part_corruption test_buffer_profile/test.py::test_buffer_profile test_buffer_profile/test.py::test_default_profile test_build_sets_from_multiple_threads/test.py::test_set test_cancel_freeze/test.py::test_cancel_backup test_cgroup_limit/test.py::test_cgroup_cpu_limit test_check_table/test.py::test_check_all_tables 'test_check_table/test.py::test_check_normal_table_corruption[]' 'test_check_table/test.py::test_check_replicated_table_corruption[-_0-.bin]' 'test_check_table/test.py::test_check_replicated_table_simple[-_0]' test_cleanup_after_start/test.py::test_old_dirs_cleanup test_cluster_all_replicas/test.py::test_cluster 'test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[one_shard_three_nodes]' 'test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[two_shards_three_nodes]' test_cluster_all_replicas/test.py::test_global_in 'test_cluster_all_replicas/test.py::test_skip_unavailable_replica[one_shard_three_nodes]' 'test_cluster_all_replicas/test.py::test_skip_unavailable_replica[two_shards_three_nodes]' test_cluster_discovery/test_password.py::test_connect_with_password test_compatibility_merge_tree_settings/test.py::test_check_projections_compatibility test_compatibility_merge_tree_settings/test.py::test_config_overrides_compatibility test_compression_nested_columns/test.py::test_nested_compression_codec test_concurrent_queries_for_all_users_restriction/test.py::test_exception_message test_concurrent_threads_soft_limit/test.py::test_concurrent_threads_soft_limit_default test_concurrent_threads_soft_limit/test.py::test_concurrent_threads_soft_limit_defined_1 test_concurrent_threads_soft_limit/test.py::test_concurrent_threads_soft_limit_defined_50 test_concurrent_threads_soft_limit/test.py::test_concurrent_threads_soft_limit_limit_reached test_concurrent_threads_soft_limit/test.py::test_use_concurrency_control_default test_concurrent_threads_soft_limit/test.py::test_use_concurrency_control_soft_limit_defined_50 test_config_substitutions/test.py::test_allow_databases test_config_substitutions/test.py::test_config test_config_substitutions/test.py::test_config_from_env_overrides test_config_substitutions/test.py::test_config_merge_from_env_overrides test_config_substitutions/test.py::test_config_multiple_zk_substitutions test_config_substitutions/test.py::test_include_config test_custom_settings/test.py::test_custom_settings -vvv -ss ENV CLICKHOUSE_LIBRARY_BRIDGE_BINARY_PATH /clickhouse-library-bridge ENV COMPOSE_HTTP_TIMEOUT 600 ENV DOCKER_MYSQL_PHP_CLIENT_TAG 88be89c1e3b6 ENV DOCKER_DOTNET_CLIENT_TAG 11de0b29a15d ENV CLICKHOUSE_TESTS_CLIENT_BIN_PATH /clickhouse ENV DOCKER_MYSQL_JS_CLIENT_TAG 41ba7c2ec2a1 ENV PATH /spark-3.3.2-bin-hadoop3/bin:/opt/gdb/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin ENV DOCKER_KERBERIZED_HADOOP_TAG latest ENV DOCKER_CHANNEL stable ENV DOCKER_CLIENT_TIMEOUT 300 ENV DOCKER_POSTGRESQL_JAVA_CLIENT_TAG a4eff5c7f4d6 ENV DOCKER_NGINX_DAV_TAG b55ac9cd7519 ENV DOCKER_MYSQL_GOLANG_CLIENT_TAG 9bec2a638e6e ENV PWD /ClickHouse/tests/integration ENV DOCKER_MYSQL_JAVA_CLIENT_TAG 766bff31cfe4 ENV CLICKHOUSE_ODBC_BRIDGE_BINARY_PATH /clickhouse-odbc-bridge ENV CLICKHOUSE_TESTS_BASE_CONFIG_DIR /clickhouse-config ENV TZ Etc/UTC ENV JAVA_PATH /usr/lib/jvm/java-11-openjdk-amd64/bin/java ENV DOCKER_BASE_TAG 8b2301119731 ENV SPARK_HOME /spark-3.3.2-bin-hadoop3 ENV LC_CTYPE C.UTF-8 ENV INTEGRATION_TESTS_RUN_ID 0 ENV WORKER_FREE_PORTS 30200 30201 30202 30203 30204 30205 30206 30207 30208 30209 30210 30211 30212 30213 30214 30215 30216 30217 30218 30219 30220 30221 30222 30223 30224 30225 30226 30227 30228 30229 30230 30231 30232 30233 30234 30235 30236 30237 30238 30239 30240 30241 30242 30243 30244 30245 30246 30247 30248 30249 ENV PYTEST_XDIST_TESTRUNUID f7495d73cd4a495b93766f09fca340b2 ENV PYTEST_XDIST_WORKER gw4 ENV PYTEST_XDIST_WORKER_COUNT 10 ENV PYTEST_CURRENT_TEST test_azure_blob_storage_zero_copy_replication/test.py::test_zero_copy_replication (setup) CLUSTER INIT base_config_dir:/clickhouse-config clickhouse_start_command: clickhouse server --config-file=/etc/clickhouse-server/{main_config_file} --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log Setup Keeper Cluster name: project_name:roottestazureblobstoragezerocopyreplication-gw4. Added instance name:node1 tag:8b2301119731 base_cmd:['docker', 'compose', '--env-file', '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw4/.env', '--project-name', 'roottestazureblobstoragezerocopyreplication-gw4', '--file', '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw4/node1/docker-compose.yml', '--file', '/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml', '--file', '/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_azurite.yml'] docker_compose_yml_dir:/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/ clickhouse_start_command: clickhouse server --config-file=/etc/clickhouse-server/{main_config_file} --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log Cluster name: project_name:roottestazureblobstoragezerocopyreplication-gw4. Added instance name:node2 tag:8b2301119731 base_cmd:['docker', 'compose', '--env-file', '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw4/.env', '--project-name', 'roottestazureblobstoragezerocopyreplication-gw4', '--file', '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw4/node1/docker-compose.yml', '--file', '/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml', '--file', '/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_azurite.yml', '--file', '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw4/node2/docker-compose.yml'] docker_compose_yml_dir:/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/ Starting cluster... Running tests in /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/test.py Cluster start called. is_up=False Docker networks for project roottestazureblobstoragezerocopyreplication-gw4 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestazureblobstoragezerocopyreplication-gw4 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestazureblobstoragezerocopyreplication-gw4 are DRIVER VOLUME NAME Cleanup called Docker networks for project roottestazureblobstoragezerocopyreplication-gw4 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestazureblobstoragezerocopyreplication-gw4 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestazureblobstoragezerocopyreplication-gw4 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestazureblobstoragezerocopyreplication-gw4-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestazureblobstoragezerocopyreplication-gw4 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:1 Volumes pruned: 1 Setup directory for instance: node1 Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/configs/config.d/config.xml', '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_gen/storage_conf.xml'] to /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw4/node1/configs/config.d Setup database dir /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw4/node1/database Setup logs dir /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw4/node1/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Setup directory for instance: node2 Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/configs/config.d/config.xml', '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_gen/storage_conf.xml'] to /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw4/node2/configs/config.d Setup database dir /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw4/node2/database Setup logs dir /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw4/node2/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'keeper_binary': '/clickhouse', 'keeper_cmd_prefix': 'clickhouse keeper', 'image': 'altinityinfra/integration-test:8b2301119731', 'user': '0', 'keeper_fs': 'bind', 'keeper_logs_dir1': '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw4/keeper1/log', 'keeper_config_dir1': '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw4/keeper1/config', 'keeper_db_dir1': '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw4/keeper1/coordination', 'keeper_logs_dir2': '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw4/keeper2/log', 'keeper_config_dir2': '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw4/keeper2/config', 'keeper_db_dir2': '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw4/keeper2/coordination', 'keeper_logs_dir3': '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw4/keeper3/log', 'keeper_config_dir3': '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw4/keeper3/config', 'keeper_db_dir3': '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw4/keeper3/coordination', 'AZURITE_PORT': '30200', 'AZURITE_STORAGE_ACCOUNT_URL': 'http://azurite1:30200/devstoreaccount1', 'AZURITE_CONNECTION_STRING': 'DefaultEndpointsProtocol=http;AccountName=devstoreaccount1;AccountKey=Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==;BlobEndpoint=http://azurite1:30200/devstoreaccount1;'} stored in /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw4/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw4/.env --project-name roottestazureblobstoragezerocopyreplication-gw4 --file /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw4/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_azurite.yml --file /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw4/node2/docker-compose.yml pull] Connection dropped: socket connection error: None Failed connecting to Zookeeper within the connection retry policy. Stderr: zoo1 Skipped - Image is already being pulled by node1 Stderr: zoo2 Skipped - Image is already being pulled by node1 Stderr: zoo3 Skipped - Image is already being pulled by node1 Stderr: node2 Skipped - Image is already being pulled by node1 Stderr: azurite1 Pulling Stderr: node1 Pulling Stderr: node1 Pulled Stderr: f18232174bc9 Pulling fs layer Stderr: cb2bde55f71f Pulling fs layer Stderr: 9d0e0719fbe0 Pulling fs layer Stderr: 6f063dbd7a5d Pulling fs layer Stderr: f9e3e3d8f042 Pulling fs layer Stderr: a39ef2f62dc8 Pulling fs layer Stderr: 9a21c6b23f0e Pulling fs layer Stderr: efeb7b313b67 Pulling fs layer Stderr: 6fef65209747 Pulling fs layer Stderr: 3d377e512a83 Pulling fs layer Stderr: f9e3e3d8f042 Waiting Stderr: a39ef2f62dc8 Waiting Stderr: 9a21c6b23f0e Waiting Stderr: efeb7b313b67 Waiting Stderr: 6fef65209747 Waiting Stderr: 3d377e512a83 Waiting Stderr: 6f063dbd7a5d Waiting Stderr: 9d0e0719fbe0 Downloading [> ] 15.58kB/1.261MB Stderr: f18232174bc9 Downloading [> ] 48.34kB/3.642MB Stderr: 9d0e0719fbe0 Downloading [============> ] 309.5kB/1.261MB Stderr: cb2bde55f71f Downloading [> ] 506.1kB/50.34MB Stderr: f18232174bc9 Downloading [========> ] 637.2kB/3.642MB Stderr: 9d0e0719fbe0 Verifying Checksum Stderr: 9d0e0719fbe0 Download complete Stderr: cb2bde55f71f Downloading [==> ] 2.538MB/50.34MB Stderr: f18232174bc9 Downloading [=====================================> ] 2.702MB/3.642MB Stderr: f18232174bc9 Verifying Checksum Stderr: f18232174bc9 Download complete Stderr: f18232174bc9 Extracting [> ] 65.54kB/3.642MB Stderr: cb2bde55f71f Downloading [=======> ] 7.109MB/50.34MB Stderr: f18232174bc9 Extracting [==================================================>] 3.642MB/3.642MB Stderr: f18232174bc9 Extracting [==================================================>] 3.642MB/3.642MB Stderr: f18232174bc9 Pull complete Stderr: cb2bde55f71f Downloading [================> ] 16.25MB/50.34MB Stderr: cb2bde55f71f Downloading [===========================> ] 27.93MB/50.34MB Stderr: cb2bde55f71f Downloading [========================================> ] 41.14MB/50.34MB Stderr: cb2bde55f71f Verifying Checksum Stderr: cb2bde55f71f Download complete Stderr: cb2bde55f71f Extracting [> ] 524.3kB/50.34MB Stderr: cb2bde55f71f Extracting [=======> ] 7.864MB/50.34MB Stderr: cb2bde55f71f Extracting [==================> ] 18.35MB/50.34MB Stderr: cb2bde55f71f Extracting [==========================> ] 26.21MB/50.34MB Stderr: cb2bde55f71f Extracting [====================================> ] 36.7MB/50.34MB Stderr: cb2bde55f71f Extracting [============================================> ] 44.56MB/50.34MB Stderr: cb2bde55f71f Extracting [==============================================> ] 47.19MB/50.34MB Stderr: cb2bde55f71f Extracting [===============================================> ] 48.23MB/50.34MB Stderr: cb2bde55f71f Extracting [================================================> ] 48.76MB/50.34MB Stderr: cb2bde55f71f Extracting [=================================================> ] 49.81MB/50.34MB Stderr: cb2bde55f71f Extracting [=================================================> ] 50.33MB/50.34MB Stderr: cb2bde55f71f Extracting [==================================================>] 50.34MB/50.34MB Stderr: cb2bde55f71f Pull complete Stderr: 9d0e0719fbe0 Extracting [=> ] 32.77kB/1.261MB Stderr: 9d0e0719fbe0 Extracting [==================================================>] 1.261MB/1.261MB Stderr: 9d0e0719fbe0 Extracting [==================================================>] 1.261MB/1.261MB Stderr: 9d0e0719fbe0 Pull complete Stderr: 6f063dbd7a5d Downloading [==================================================>] 446B/446B Stderr: 6f063dbd7a5d Verifying Checksum Stderr: 6f063dbd7a5d Download complete Stderr: 6f063dbd7a5d Extracting [==================================================>] 446B/446B Stderr: 6f063dbd7a5d Extracting [==================================================>] 446B/446B Stderr: 6f063dbd7a5d Pull complete Stderr: f9e3e3d8f042 Downloading [==================================================>] 116B/116B Stderr: f9e3e3d8f042 Verifying Checksum Stderr: f9e3e3d8f042 Download complete Stderr: f9e3e3d8f042 Extracting [==================================================>] 116B/116B Stderr: f9e3e3d8f042 Extracting [==================================================>] 116B/116B Stderr: a39ef2f62dc8 Downloading [> ] 3.29kB/209.4kB Stderr: f9e3e3d8f042 Pull complete Stderr: a39ef2f62dc8 Downloading [==================================================>] 209.4kB/209.4kB Stderr: a39ef2f62dc8 Verifying Checksum Stderr: a39ef2f62dc8 Download complete Stderr: a39ef2f62dc8 Extracting [=======> ] 32.77kB/209.4kB Stderr: a39ef2f62dc8 Extracting [==================================================>] 209.4kB/209.4kB Stderr: a39ef2f62dc8 Extracting [==================================================>] 209.4kB/209.4kB Stderr: a39ef2f62dc8 Pull complete Stderr: 9a21c6b23f0e Downloading [> ] 15.58kB/794kB Stderr: efeb7b313b67 Downloading [=> ] 15.58kB/458.8kB Stderr: 9a21c6b23f0e Downloading [======================> ] 358.7kB/794kB Stderr: efeb7b313b67 Downloading [=================================> ] 309.5kB/458.8kB Stderr: 6fef65209747 Downloading [> ] 375.1kB/36.34MB Stderr: efeb7b313b67 Verifying Checksum Stderr: efeb7b313b67 Download complete Stderr: 9a21c6b23f0e Downloading [==================================================>] 794kB/794kB Stderr: 9a21c6b23f0e Download complete Stderr: 9a21c6b23f0e Extracting [==> ] 32.77kB/794kB Stderr: 6fef65209747 Downloading [===> ] 2.259MB/36.34MB Stderr: 9a21c6b23f0e Extracting [==============================> ] 491.5kB/794kB Stderr: 6fef65209747 Downloading [=========> ] 6.781MB/36.34MB Stderr: 9a21c6b23f0e Extracting [=============================================> ] 720.9kB/794kB Stderr: 9a21c6b23f0e Extracting [==================================================>] 794kB/794kB Stderr: 6fef65209747 Downloading [=====================> ] 15.83MB/36.34MB Stderr: 9a21c6b23f0e Pull complete Stderr: efeb7b313b67 Extracting [===> ] 32.77kB/458.8kB Stderr: efeb7b313b67 Extracting [==================================================>] 458.8kB/458.8kB Stderr: efeb7b313b67 Extracting [==================================================>] 458.8kB/458.8kB Stderr: 6fef65209747 Downloading [========================================> ] 29.39MB/36.34MB Stderr: efeb7b313b67 Pull complete Stderr: 6fef65209747 Verifying Checksum Stderr: 6fef65209747 Download complete Stderr: 6fef65209747 Extracting [> ] 393.2kB/36.34MB Stderr: 6fef65209747 Extracting [=> ] 786.4kB/36.34MB Stderr: 6fef65209747 Extracting [=> ] 1.18MB/36.34MB Stderr: 6fef65209747 Extracting [==> ] 1.573MB/36.34MB Stderr: 6fef65209747 Extracting [===> ] 2.359MB/36.34MB Stderr: 6fef65209747 Extracting [===> ] 2.753MB/36.34MB Stderr: 6fef65209747 Extracting [=====> ] 3.932MB/36.34MB Stderr: 6fef65209747 Extracting [=======> ] 5.112MB/36.34MB Stderr: 6fef65209747 Extracting [========> ] 5.898MB/36.34MB Stderr: 6fef65209747 Extracting [========> ] 6.291MB/36.34MB Stderr: 6fef65209747 Extracting [=========> ] 7.078MB/36.34MB Stderr: 6fef65209747 Extracting [==========> ] 7.471MB/36.34MB Stderr: 6fef65209747 Extracting [==========> ] 7.864MB/36.34MB Stderr: 6fef65209747 Extracting [===========> ] 8.258MB/36.34MB Stderr: 6fef65209747 Extracting [===========> ] 8.651MB/36.34MB Stderr: 6fef65209747 Extracting [============> ] 9.044MB/36.34MB Stderr: 6fef65209747 Extracting [=============> ] 9.83MB/36.34MB Stderr: 6fef65209747 Extracting [=================> ] 12.98MB/36.34MB Stderr: 6fef65209747 Extracting [===================> ] 14.16MB/36.34MB Stderr: 6fef65209747 Extracting [====================> ] 14.94MB/36.34MB Stderr: 6fef65209747 Extracting [=====================> ] 15.73MB/36.34MB Stderr: 6fef65209747 Extracting [======================> ] 16.12MB/36.34MB Stderr: 6fef65209747 Extracting [=======================> ] 17.3MB/36.34MB Stderr: 6fef65209747 Extracting [============================> ] 20.45MB/36.34MB Stderr: 6fef65209747 Extracting [======================================> ] 28.31MB/36.34MB Stderr: 6fef65209747 Extracting [==========================================> ] 31.06MB/36.34MB Stderr: 6fef65209747 Extracting [==================================================>] 36.34MB/36.34MB Stderr: 6fef65209747 Pull complete Stderr: 3d377e512a83 Downloading [==================================================>] 2.862kB/2.862kB Stderr: 3d377e512a83 Verifying Checksum Stderr: 3d377e512a83 Download complete Stderr: 3d377e512a83 Extracting [==================================================>] 2.862kB/2.862kB Stderr: 3d377e512a83 Extracting [==================================================>] 2.862kB/2.862kB Stderr: 3d377e512a83 Pull complete Stderr: azurite1 Pulled Setup ZooKeeper Creating internal ZooKeeper dirs: ['/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw4/keeper1/log', '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw4/keeper1/config', '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw4/keeper1/coordination', '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw4/keeper2/log', '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw4/keeper2/config', '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw4/keeper2/coordination', '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw4/keeper3/log', '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw4/keeper3/config', '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw4/keeper3/coordination'] Command:[docker compose --project-name roottestazureblobstoragezerocopyreplication-gw4 --env-file /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw4/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --verbose up -d] Stderr:time="2025-04-02T03:42:11Z" level=trace msg="Docker Desktop integration not enabled" Stderr: Network roottestazureblobstoragezerocopyreplication-gw4_default Creating Stderr: Network roottestazureblobstoragezerocopyreplication-gw4_default Created Stderr: Container roottestazureblobstoragezerocopyreplication-gw4-zoo2-1 Creating Stderr: Container roottestazureblobstoragezerocopyreplication-gw4-zoo1-1 Creating Stderr: Container roottestazureblobstoragezerocopyreplication-gw4-zoo3-1 Creating Stderr: Container roottestazureblobstoragezerocopyreplication-gw4-zoo2-1 Created Stderr: Container roottestazureblobstoragezerocopyreplication-gw4-zoo3-1 Created Stderr: Container roottestazureblobstoragezerocopyreplication-gw4-zoo1-1 Created Stderr: Container roottestazureblobstoragezerocopyreplication-gw4-zoo1-1 Starting Stderr: Container roottestazureblobstoragezerocopyreplication-gw4-zoo2-1 Starting Stderr: Container roottestazureblobstoragezerocopyreplication-gw4-zoo3-1 Starting Stderr: Container roottestazureblobstoragezerocopyreplication-gw4-zoo1-1 Started Stderr: Container roottestazureblobstoragezerocopyreplication-gw4-zoo2-1 Started Stderr: Container roottestazureblobstoragezerocopyreplication-gw4-zoo3-1 Started Stderr:time="2025-04-02T03:42:11Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T03:42:11Z" level=debug msg="otel error" error="" Wait ZooKeeper to start get_instance_ip instance_name=zoo1 http://localhost:None "GET /v1.46/containers/roottestazureblobstoragezerocopyreplication-gw4-zoo1-1/json HTTP/1.1" 200 None get_kazoo_client: zoo1, ip:172.16.1.2, port:2181, use_ssl:False Connecting to 172.16.1.2(172.16.1.2):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Connecting to 172.16.1.2(172.16.1.2):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Connecting to 172.16.1.2(172.16.1.2):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Connecting to 172.16.1.2(172.16.1.2):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Connecting to 172.16.1.2(172.16.1.2):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Connecting to 172.16.1.2(172.16.1.2):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Connecting to 172.16.1.2(172.16.1.2):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED get_instance_ip instance_name=zoo2 http://localhost:None "GET /v1.46/containers/roottestazureblobstoragezerocopyreplication-gw4-zoo2-1/json HTTP/1.1" 200 None get_kazoo_client: zoo2, ip:172.16.1.4, port:2181, use_ssl:False Connecting to 172.16.1.4(172.16.1.4):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED get_instance_ip instance_name=zoo3 http://localhost:None "GET /v1.46/containers/roottestazureblobstoragezerocopyreplication-gw4-zoo3-1/json HTTP/1.1" 200 None get_kazoo_client: zoo3, ip:172.16.1.3, port:2181, use_ssl:False Connecting to 172.16.1.3(172.16.1.3):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED All instances of ZooKeeper started: ('zoo1', 'zoo2', 'zoo3') Trying to create Azurite instance by command docker compose --project-name roottestazureblobstoragezerocopyreplication-gw4 --env-file /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw4/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_azurite.yml --verbose up -d Command:[docker compose --project-name roottestazureblobstoragezerocopyreplication-gw4 --env-file /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw4/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_azurite.yml --verbose up -d] Stderr:time="2025-04-02T03:42:16Z" level=trace msg="Docker Desktop integration not enabled" Stderr: Volume "roottestazureblobstoragezerocopyreplication-gw4_data1-1" Creating Stderr: Volume "roottestazureblobstoragezerocopyreplication-gw4_data1-1" Created Stderr:time="2025-04-02T03:42:16Z" level=warning msg="Found orphan containers ([roottestazureblobstoragezerocopyreplication-gw4-zoo1-1 roottestazureblobstoragezerocopyreplication-gw4-zoo3-1 roottestazureblobstoragezerocopyreplication-gw4-zoo2-1]) for this project. If you removed or renamed this service in your compose file, you can run this command with the --remove-orphans flag to clean it up." Stderr: Container roottestazureblobstoragezerocopyreplication-gw4-azurite1-1 Creating Stderr: Container roottestazureblobstoragezerocopyreplication-gw4-azurite1-1 Created Stderr: Container roottestazureblobstoragezerocopyreplication-gw4-azurite1-1 Starting Stderr: Container roottestazureblobstoragezerocopyreplication-gw4-azurite1-1 Started Stderr:time="2025-04-02T03:42:16Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T03:42:16Z" level=debug msg="otel error" error="" Trying to connect to Azurite Request URL: 'http://127.0.0.1:30200/devstoreaccount1/?restype=REDACTED&comp=REDACTED' Request method: 'GET' Request headers: 'x-ms-version': 'REDACTED' 'Accept': 'application/xml' 'User-Agent': 'azsdk-python-storage-blob/12.19.0 Python/3.10.12 (Linux-5.15.0-130-generic-x86_64-with-glibc2.35)' 'x-ms-date': 'REDACTED' 'x-ms-client-request-id': '795a16e6-0f74-11f0-9258-0242ac110002' 'Authorization': 'REDACTED' No body was attached to the request Starting new HTTP connection (1): 127.0.0.1:30200 http://127.0.0.1:30200 "GET /devstoreaccount1/?restype=account&comp=properties HTTP/1.1" 200 0 Response status: 200 Response headers: 'Server': 'Azurite-Blob/3.34.0' 'x-ms-client-request-id': '795a16e6-0f74-11f0-9258-0242ac110002' 'x-ms-request-id': '66c4e363-0afb-449b-a4a9-7b4b0cb00384' 'x-ms-version': 'REDACTED' 'date': 'Wed, 02 Apr 2025 03:42:17 GMT' 'x-ms-sku-name': 'REDACTED' 'x-ms-account-kind': 'REDACTED' 'x-ms-is-hns-enabled': 'REDACTED' 'Connection': 'keep-alive' 'Keep-Alive': 'REDACTED' 'Content-Length': '0' {'client_request_id': '795a16e6-0f74-11f0-9258-0242ac110002', 'request_id': '66c4e363-0afb-449b-a4a9-7b4b0cb00384', 'version': '2025-05-05', 'date': datetime.datetime(2025, 4, 2, 3, 42, 17, tzinfo=datetime.timezone.utc), 'sku_name': 'Standard_RAGRS', 'account_kind': 'StorageV2', 'is_hns_enabled': False} Request URL: 'http://127.0.0.1:30200/devstoreaccount1/?comp=REDACTED&prefix=REDACTED&include=REDACTED' Request method: 'GET' Request headers: 'x-ms-version': 'REDACTED' 'Accept': 'application/xml' 'User-Agent': 'azsdk-python-storage-blob/12.19.0 Python/3.10.12 (Linux-5.15.0-130-generic-x86_64-with-glibc2.35)' 'x-ms-date': 'REDACTED' 'x-ms-client-request-id': '79642b90-0f74-11f0-9258-0242ac110002' 'Authorization': 'REDACTED' No body was attached to the request http://127.0.0.1:30200 "GET /devstoreaccount1/?comp=list&prefix=azurite-container&include= HTTP/1.1" 200 None Response status: 200 Response headers: 'Server': 'Azurite-Blob/3.34.0' 'x-ms-client-request-id': '79642b90-0f74-11f0-9258-0242ac110002' 'x-ms-request-id': '08121a06-ac86-40ec-9afb-8e5081d15180' 'x-ms-version': 'REDACTED' 'content-type': 'application/xml' 'Date': 'Wed, 02 Apr 2025 03:42:17 GMT' 'Connection': 'keep-alive' 'Keep-Alive': 'REDACTED' 'Transfer-Encoding': 'chunked' Request URL: 'http://127.0.0.1:30200/devstoreaccount1/azurite-container?restype=REDACTED' Request method: 'GET' Request headers: 'x-ms-version': 'REDACTED' 'Accept': 'application/xml' 'User-Agent': 'azsdk-python-storage-blob/12.19.0 Python/3.10.12 (Linux-5.15.0-130-generic-x86_64-with-glibc2.35)' 'x-ms-date': 'REDACTED' 'x-ms-client-request-id': '79663b10-0f74-11f0-9258-0242ac110002' 'Authorization': 'REDACTED' No body was attached to the request http://127.0.0.1:30200 "GET /devstoreaccount1/azurite-container?restype=container HTTP/1.1" 404 None Response status: 404 Response headers: 'Server': 'Azurite-Blob/3.34.0' 'x-ms-error-code': 'ContainerNotFound' 'x-ms-request-id': 'b47c3678-00d3-469c-a0bc-79f79099bdfd' 'content-type': 'application/xml' 'Date': 'Wed, 02 Apr 2025 03:42:17 GMT' 'Connection': 'keep-alive' 'Keep-Alive': 'REDACTED' 'Transfer-Encoding': 'chunked' azurite container 'azurite-container' doesn't exist, creating it Request URL: 'http://127.0.0.1:30200/devstoreaccount1/azurite-container?restype=REDACTED' Request method: 'PUT' Request headers: 'x-ms-version': 'REDACTED' 'Accept': 'application/xml' 'User-Agent': 'azsdk-python-storage-blob/12.19.0 Python/3.10.12 (Linux-5.15.0-130-generic-x86_64-with-glibc2.35)' 'x-ms-date': 'REDACTED' 'x-ms-client-request-id': '7967c8ae-0f74-11f0-9258-0242ac110002' 'Authorization': 'REDACTED' No body was attached to the request http://127.0.0.1:30200 "PUT /devstoreaccount1/azurite-container?restype=container HTTP/1.1" 201 0 Response status: 201 Response headers: 'Server': 'Azurite-Blob/3.34.0' 'etag': '"0x1FB865B0CB42400"' 'last-modified': 'Wed, 02 Apr 2025 03:42:17 GMT' 'x-ms-client-request-id': '7967c8ae-0f74-11f0-9258-0242ac110002' 'x-ms-request-id': '1874a17c-2a62-4edc-9276-4554ef90e10c' 'x-ms-version': 'REDACTED' 'Date': 'Wed, 02 Apr 2025 03:42:17 GMT' 'Connection': 'keep-alive' 'Keep-Alive': 'REDACTED' 'Content-Length': '0' ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw4/.env --project-name roottestazureblobstoragezerocopyreplication-gw4 --file /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw4/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_azurite.yml --file /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw4/node2/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw4/.env --project-name roottestazureblobstoragezerocopyreplication-gw4 --file /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw4/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_azurite.yml --file /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw4/node2/docker-compose.yml up -d --no-recreate] Stderr: Container roottestazureblobstoragezerocopyreplication-gw4-zoo1-1 Running Stderr: Container roottestazureblobstoragezerocopyreplication-gw4-zoo3-1 Running Stderr: Container roottestazureblobstoragezerocopyreplication-gw4-zoo2-1 Running Stderr: Container roottestazureblobstoragezerocopyreplication-gw4-azurite1-1 Running Stderr: Container roottestazureblobstoragezerocopyreplication-gw4-node1-1 Creating Stderr: Container roottestazureblobstoragezerocopyreplication-gw4-node2-1 Creating Stderr: Container roottestazureblobstoragezerocopyreplication-gw4-node1-1 Created Stderr: Container roottestazureblobstoragezerocopyreplication-gw4-node2-1 Created Stderr: Container roottestazureblobstoragezerocopyreplication-gw4-node2-1 Starting Stderr: Container roottestazureblobstoragezerocopyreplication-gw4-node1-1 Starting Stderr: Container roottestazureblobstoragezerocopyreplication-gw4-node1-1 Started Stderr: Container roottestazureblobstoragezerocopyreplication-gw4-node2-1 Started ClickHouse instance created get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottestazureblobstoragezerocopyreplication-gw4-node1-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottestazureblobstoragezerocopyreplication-gw4-node1-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node1, ip: 172.16.1.7... http://localhost:None "GET /v1.46/containers/roottestazureblobstoragezerocopyreplication-gw4-node1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/e11685ba65797791b77da415a5632c401a362929c6b6b621461d1501ce031289/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/e11685ba65797791b77da415a5632c401a362929c6b6b621461d1501ce031289/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/e11685ba65797791b77da415a5632c401a362929c6b6b621461d1501ce031289/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/e11685ba65797791b77da415a5632c401a362929c6b6b621461d1501ce031289/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/e11685ba65797791b77da415a5632c401a362929c6b6b621461d1501ce031289/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/e11685ba65797791b77da415a5632c401a362929c6b6b621461d1501ce031289/json HTTP/1.1" 200 None ClickHouse node1 started get_instance_ip instance_name=node2 http://localhost:None "GET /v1.46/containers/roottestazureblobstoragezerocopyreplication-gw4-node2-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node2 http://localhost:None "GET /v1.46/containers/roottestazureblobstoragezerocopyreplication-gw4-node2-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node2, ip: 172.16.1.6... http://localhost:None "GET /v1.46/containers/roottestazureblobstoragezerocopyreplication-gw4-node2-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/ee64c02c4d7f751aa2707bfb7e4a7da2031be08463c4e8cf069e1739c72ea5f0/json HTTP/1.1" 200 None ClickHouse node2 started Cluster started Executing query CREATE TABLE blob_storage_table ON CLUSTER test_cluster ( id Int64, data String ) ENGINE=ReplicatedMergeTree('/clickhouse/tables/blob_storage_table', '{replica}') ORDER BY id SETTINGS storage_policy='blob_storage_policy',old_parts_lifetime=1 on node1 Executing query SELECT COUNT(*) FROM blob_storage_table FORMAT Values on node1 Executing query INSERT INTO blob_storage_table VALUES (0,'data'),(1,'data') on node1 Executing query SYSTEM SYNC REPLICA blob_storage_table on node2 Executing query SELECT * FROM blob_storage_table order by id FORMAT Values on node1 Executing query SELECT * FROM blob_storage_table order by id FORMAT Values on node2 Request URL: 'http://127.0.0.1:30200/devstoreaccount1/cont?restype=REDACTED&comp=REDACTED' Request method: 'GET' Request headers: 'x-ms-version': 'REDACTED' 'Accept': 'application/xml' 'User-Agent': 'azsdk-python-storage-blob/12.19.0 Python/3.10.12 (Linux-5.15.0-130-generic-x86_64-with-glibc2.35)' 'x-ms-date': 'REDACTED' 'x-ms-client-request-id': '7a537eac-0f74-11f0-9258-0242ac110002' 'Authorization': 'REDACTED' No body was attached to the request http://127.0.0.1:30200 "GET /devstoreaccount1/cont?restype=container&comp=list HTTP/1.1" 200 None Response status: 200 Response headers: 'Server': 'Azurite-Blob/3.34.0' 'content-type': 'application/xml' 'x-ms-client-request-id': '7a537eac-0f74-11f0-9258-0242ac110002' 'x-ms-request-id': '5b6c2d59-79af-499f-a517-eec22bb9c82d' 'x-ms-version': 'REDACTED' 'date': 'Wed, 02 Apr 2025 03:42:19 GMT' 'Connection': 'keep-alive' 'Keep-Alive': 'REDACTED' 'Transfer-Encoding': 'chunked' Executing query INSERT INTO blob_storage_table VALUES (2,'data'),(3,'data') on node2 Executing query SYSTEM SYNC REPLICA blob_storage_table on node1 Executing query SELECT * FROM blob_storage_table order by id FORMAT Values on node2 Executing query SELECT * FROM blob_storage_table order by id FORMAT Values on node1 Request URL: 'http://127.0.0.1:30200/devstoreaccount1/cont?restype=REDACTED&comp=REDACTED' Request method: 'GET' Request headers: 'x-ms-version': 'REDACTED' 'Accept': 'application/xml' 'User-Agent': 'azsdk-python-storage-blob/12.19.0 Python/3.10.12 (Linux-5.15.0-130-generic-x86_64-with-glibc2.35)' 'x-ms-date': 'REDACTED' 'x-ms-client-request-id': '7a8c7234-0f74-11f0-9258-0242ac110002' 'Authorization': 'REDACTED' No body was attached to the request http://127.0.0.1:30200 "GET /devstoreaccount1/cont?restype=container&comp=list HTTP/1.1" 200 None Response status: 200 Response headers: 'Server': 'Azurite-Blob/3.34.0' 'content-type': 'application/xml' 'x-ms-client-request-id': '7a8c7234-0f74-11f0-9258-0242ac110002' 'x-ms-request-id': '5f2a554c-dcf5-4b3e-afc7-217593af531f' 'x-ms-version': 'REDACTED' 'date': 'Wed, 02 Apr 2025 03:42:19 GMT' 'Connection': 'keep-alive' 'Keep-Alive': 'REDACTED' 'Transfer-Encoding': 'chunked' Executing query DROP TABLE blob_storage_table ON CLUSTER test_cluster SYNC on node1 Command:[docker compose --env-file /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw4/.env --project-name roottestazureblobstoragezerocopyreplication-gw4 --file /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw4/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_azurite.yml --file /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw4/node2/docker-compose.yml stop --timeout 20] [gw4] PASSED test_azure_blob_storage_zero_copy_replication/test.py::test_zero_copy_replication Stderr: Container roottestazureblobstoragezerocopyreplication-gw4-node1-1 Stopping Stderr: Container roottestazureblobstoragezerocopyreplication-gw4-node2-1 Stopping Stderr: Container roottestazureblobstoragezerocopyreplication-gw4-node1-1 Stopped Stderr: Container roottestazureblobstoragezerocopyreplication-gw4-node2-1 Stopped Stderr: Container roottestazureblobstoragezerocopyreplication-gw4-zoo1-1 Stopping Stderr: Container roottestazureblobstoragezerocopyreplication-gw4-zoo3-1 Stopping Stderr: Container roottestazureblobstoragezerocopyreplication-gw4-zoo2-1 Stopping Stderr: Container roottestazureblobstoragezerocopyreplication-gw4-azurite1-1 Stopping Stderr: Container roottestazureblobstoragezerocopyreplication-gw4-zoo3-1 Stopped Stderr: Container roottestazureblobstoragezerocopyreplication-gw4-zoo1-1 Stopped Stderr: Container roottestazureblobstoragezerocopyreplication-gw4-zoo2-1 Stopped Stderr: Container roottestazureblobstoragezerocopyreplication-gw4-azurite1-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw4/node1/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw4/node1/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw4/node2/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw4/node2/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw4/.env --project-name roottestazureblobstoragezerocopyreplication-gw4 --file /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw4/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_azurite.yml --file /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw4/node2/docker-compose.yml down --volumes] Stderr: Container roottestazureblobstoragezerocopyreplication-gw4-node1-1 Stopping Stderr: Container roottestazureblobstoragezerocopyreplication-gw4-node2-1 Stopping Stderr: Container roottestazureblobstoragezerocopyreplication-gw4-node1-1 Stopped Stderr: Container roottestazureblobstoragezerocopyreplication-gw4-node1-1 Removing Stderr: Container roottestazureblobstoragezerocopyreplication-gw4-node2-1 Stopped Stderr: Container roottestazureblobstoragezerocopyreplication-gw4-node2-1 Removing Stderr: Container roottestazureblobstoragezerocopyreplication-gw4-node1-1 Removed Stderr: Container roottestazureblobstoragezerocopyreplication-gw4-node2-1 Removed Stderr: Container roottestazureblobstoragezerocopyreplication-gw4-zoo3-1 Stopping Stderr: Container roottestazureblobstoragezerocopyreplication-gw4-zoo2-1 Stopping Stderr: Container roottestazureblobstoragezerocopyreplication-gw4-zoo1-1 Stopping Stderr: Container roottestazureblobstoragezerocopyreplication-gw4-azurite1-1 Stopping Stderr: Container roottestazureblobstoragezerocopyreplication-gw4-zoo3-1 Stopped Stderr: Container roottestazureblobstoragezerocopyreplication-gw4-zoo3-1 Removing Stderr: Container roottestazureblobstoragezerocopyreplication-gw4-zoo1-1 Stopped Stderr: Container roottestazureblobstoragezerocopyreplication-gw4-zoo1-1 Removing Stderr: Container roottestazureblobstoragezerocopyreplication-gw4-zoo2-1 Stopped Stderr: Container roottestazureblobstoragezerocopyreplication-gw4-zoo2-1 Removing Stderr: Container roottestazureblobstoragezerocopyreplication-gw4-azurite1-1 Stopped Stderr: Container roottestazureblobstoragezerocopyreplication-gw4-azurite1-1 Removing Stderr: Container roottestazureblobstoragezerocopyreplication-gw4-zoo3-1 Removed Stderr: Container roottestazureblobstoragezerocopyreplication-gw4-zoo2-1 Removed Stderr: Container roottestazureblobstoragezerocopyreplication-gw4-zoo1-1 Removed Stderr: Container roottestazureblobstoragezerocopyreplication-gw4-azurite1-1 Removed Stderr: Volume roottestazureblobstoragezerocopyreplication-gw4_data1-1 Removing Stderr: Network roottestazureblobstoragezerocopyreplication-gw4_default Removing Stderr: Volume roottestazureblobstoragezerocopyreplication-gw4_data1-1 Removed Stderr: Network roottestazureblobstoragezerocopyreplication-gw4_default Removed Cleanup called Docker networks for project roottestazureblobstoragezerocopyreplication-gw4 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestazureblobstoragezerocopyreplication-gw4 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestazureblobstoragezerocopyreplication-gw4 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestazureblobstoragezerocopyreplication-gw4-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestazureblobstoragezerocopyreplication-gw4 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:1 Volumes pruned: 1 ============================== slowest durations =============================== 305.75s call test_check_table/test.py::test_check_replicated_table_corruption[-_0-.bin] 174.54s setup test_azure_blob_storage_zero_copy_replication/test.py::test_zero_copy_replication 100.22s call test_backward_compatibility/test_functions.py::test_string_functions 90.15s call test_backward_compatibility/test_convert_ordinary.py::test_convert_ordinary_to_atomic 39.66s call test_backup_restore_new/test_shutdown_wait_backup.py::test_shutdown_wait_backup 28.46s call test_cluster_all_replicas/test.py::test_skip_unavailable_replica[two_shards_three_nodes] 23.88s setup test_cluster_all_replicas/test.py::test_cluster 20.70s teardown test_azure_blob_storage_zero_copy_replication/test.py::test_zero_copy_replication 19.52s setup test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-False] 19.51s teardown test_broken_part_during_merge/test.py::test_merge_and_part_corruption 19.08s setup test_backward_compatibility/test_insert_profile_events.py::test_new_client_compatible 18.72s setup test_config_substitutions/test.py::test_allow_databases 18.51s setup test_allowed_client_hosts/test.py::test_allowed_host 17.68s teardown test_compatibility_merge_tree_settings/test.py::test_config_overrides_compatibility 17.47s setup test_backward_compatibility/test_convert_ordinary.py::test_convert_ordinary_to_atomic 17.28s setup test_alter_on_mixed_type_cluster/test.py::test_alter_on_cluter_non_replicated 16.80s setup test_backward_compatibility/test_vertical_merges_from_compact_parts.py::test_vertical_merges_from_compact_parts 16.47s setup test_async_connect_to_multiple_ips/test.py::test 16.21s teardown test_alternative_keeper_config/test.py::test_create_insert 16.19s setup test_check_table/test.py::test_check_all_tables 16.10s setup test_compression_nested_columns/test.py::test_nested_compression_codec 15.92s setup test_backward_compatibility/test_functions.py::test_aggregate_states 15.75s setup test_broken_part_during_merge/test.py::test_merge_and_part_corruption 15.50s call test_backward_compatibility/test_vertical_merges_from_compact_parts.py::test_vertical_merges_from_compact_parts 15.24s teardown test_config_substitutions/test.py::test_include_config 14.57s setup test_backup_restore/test.py::test_attach_partition 14.34s call test_backward_compatibility/test_functions.py::test_aggregate_states 13.57s setup test_access_for_functions/test.py::test_access_rights_for_function 13.55s setup test_attach_partition_with_large_destination/test.py::test_attach_partition_with_large_destination[Atomic] 13.46s setup test_async_metrics_in_cgroup/test.py::test_normalized_user_cpu 13.12s call test_backup_restore_on_cluster/test.py::test_shutdown_waits_for_backup 13.09s setup test_cleanup_after_start/test.py::test_old_dirs_cleanup 13.02s teardown test_cluster_all_replicas/test.py::test_skip_unavailable_replica[two_shards_three_nodes] 13.01s setup test_concurrent_threads_soft_limit/test.py::test_concurrent_threads_soft_limit_default 12.76s call test_cgroup_limit/test.py::test_cgroup_cpu_limit 12.75s setup test_attach_table_normalizer/test.py::test_attach_substr 12.35s call test_cancel_freeze/test.py::test_cancel_backup 12.19s setup test_build_sets_from_multiple_threads/test.py::test_set 12.05s teardown test_backup_restore_on_cluster/test.py::test_tables_dependency 12.00s setup test_alter_codec/test.py::test_alter_codec_index 11.89s setup test_cancel_freeze/test.py::test_cancel_backup 11.44s setup test_custom_settings/test.py::test_custom_settings 11.37s setup test_alternative_keeper_config/test.py::test_create_insert 11.34s setup test_cluster_discovery/test_password.py::test_connect_with_password 11.31s call test_async_metrics_in_cgroup/test.py::test_user_cpu_accounting 11.07s teardown test_alter_on_mixed_type_cluster/test.py::test_alter_replicated_on_cluster 10.91s call test_async_metrics_in_cgroup/test.py::test_system_wide_metrics 10.73s call test_async_metrics_in_cgroup/test.py::test_normalized_user_cpu 10.59s setup test_compatibility_merge_tree_settings/test.py::test_check_projections_compatibility 10.31s setup test_aggregation_memory_efficient/test.py::test_remote 9.71s setup test_backward_compatibility/test_normalized_count_comparison.py::test_select_aggregate_alias_column 9.46s call test_check_table/test.py::test_check_normal_table_corruption[] 8.54s teardown test_check_table/test.py::test_check_replicated_table_simple[-_0] 7.88s setup test_authentication/test.py::test_authentication_fail 7.87s setup test_backward_compatibility/test_cte_distributed.py::test_cte_distributed 7.70s call test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[one_shard_three_nodes] 7.35s setup test_backward_compatibility/test_select_aggregate_alias_column.py::test_select_aggregate_alias_column 6.32s teardown test_async_connect_to_multiple_ips/test.py::test 6.27s call test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[two_shards_three_nodes] 6.07s call test_cluster_all_replicas/test.py::test_skip_unavailable_replica[one_shard_three_nodes] 5.89s teardown test_cancel_freeze/test.py::test_cancel_backup 5.79s teardown test_compression_nested_columns/test.py::test_nested_compression_codec 5.64s teardown test_build_sets_from_multiple_threads/test.py::test_set 5.59s call test_backup_restore/test.py::test_attach_partition 5.50s setup test_backup_restore_new/test_shutdown_wait_backup.py::test_shutdown_wait_backup 5.25s call test_check_table/test.py::test_check_all_tables 5.22s teardown test_concurrent_threads_soft_limit/test.py::test_use_concurrency_control_soft_limit_defined_50 5.21s setup test_buffer_profile/test.py::test_buffer_profile 5.08s call test_backup_restore/test.py::test_replace_partition 5.04s call test_cleanup_after_start/test.py::test_old_dirs_cleanup 5.03s setup test_concurrent_queries_for_all_users_restriction/test.py::test_exception_message 4.95s call test_access_for_functions/test.py::test_ignore_obsolete_grant_on_database 4.83s call test_backup_restore/test.py::test_restore 4.66s teardown test_backward_compatibility/test_cte_distributed.py::test_cte_distributed 4.63s teardown test_allowed_client_hosts/test.py::test_denied_host 4.59s teardown test_custom_settings/test.py::test_custom_settings 4.56s teardown test_alter_codec/test.py::test_alter_codec_pk 4.55s teardown test_attach_partition_with_large_destination/test.py::test_attach_partition_with_large_destination[Ordinary] 4.53s teardown test_backward_compatibility/test_select_aggregate_alias_column.py::test_select_aggregate_alias_column 4.47s teardown test_concurrent_queries_for_all_users_restriction/test.py::test_exception_message 4.42s teardown test_buffer_profile/test.py::test_default_profile 4.36s teardown test_backward_compatibility/test_insert_profile_events.py::test_old_client_compatible 4.30s call test_config_substitutions/test.py::test_config_multiple_zk_substitutions 4.22s teardown test_backward_compatibility/test_functions.py::test_string_functions 4.02s teardown test_backward_compatibility/test_normalized_count_comparison.py::test_select_aggregate_alias_column 3.93s call test_backup_restore_on_cluster/test.py::test_mutation 3.92s call test_async_connect_to_multiple_ips/test.py::test 3.92s teardown test_backup_restore/test.py::test_restore 3.70s teardown test_authentication/test.py::test_authentication_pass 3.67s call test_backup_restore_on_cluster/test.py::test_replicated_database_async 3.64s call test_check_table/test.py::test_check_replicated_table_simple[-_0] 3.49s teardown test_aggregation_memory_efficient/test.py::test_remote 3.29s call test_build_sets_from_multiple_threads/test.py::test_set 2.87s call test_attach_table_normalizer/test.py::test_attach_substr_restart 2.85s call test_backup_restore_on_cluster/test.py::test_tables_dependency 2.62s call test_backup_restore_on_cluster/test.py::test_replicated_database 2.61s call test_backup_restore_on_cluster/test.py::test_keeper_value_max_size 2.59s teardown test_backward_compatibility/test_vertical_merges_from_compact_parts.py::test_vertical_merges_from_compact_parts 2.27s teardown test_cleanup_after_start/test.py::test_old_dirs_cleanup 2.26s teardown test_cluster_discovery/test_password.py::test_connect_with_password 2.25s call test_broken_part_during_merge/test.py::test_merge_and_part_corruption 2.17s call test_backup_restore_on_cluster/test.py::test_replicated_database_compare_parts 2.15s teardown test_backward_compatibility/test_convert_ordinary.py::test_convert_ordinary_to_atomic 2.13s call test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_def 2.12s call test_alter_on_mixed_type_cluster/test.py::test_alter_on_cluter_non_replicated 2.05s call test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[database] 2.03s call test_backup_restore_on_cluster/test.py::test_table_in_replicated_database_with_not_synced_def 2.00s call test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_merge 2.00s call test_alter_codec/test.py::test_alter_codec_index 1.95s call test_backup_restore_on_cluster/test.py::test_required_privileges 1.94s call test_backup_restore_on_cluster/test.py::test_replicated_table 1.93s call test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-True] 1.73s call test_alter_on_mixed_type_cluster/test.py::test_alter_replicated_on_cluster 1.60s call test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[uuid] 1.58s call test_concurrent_queries_for_all_users_restriction/test.py::test_exception_message 1.57s call test_compatibility_merge_tree_settings/test.py::test_check_projections_compatibility 1.49s teardown test_attach_table_normalizer/test.py::test_attach_substr_restart 1.48s call test_backup_restore_on_cluster/test.py::test_backup_restore_on_single_replica 1.48s call test_backup_restore_on_cluster/test.py::test_replicated_table_with_uuid_in_zkpath 1.45s teardown test_async_metrics_in_cgroup/test.py::test_user_cpu_accounting 1.45s call test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_insert 1.43s call test_backup_restore_on_cluster/test.py::test_system_functions 1.41s teardown test_backup_restore_new/test_shutdown_wait_backup.py::test_shutdown_wait_backup 1.40s call test_concurrent_threads_soft_limit/test.py::test_concurrent_threads_soft_limit_limit_reached 1.39s call test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-False] 1.39s teardown test_access_for_functions/test.py::test_ignore_obsolete_grant_on_database 1.38s call test_backup_restore_on_cluster/test.py::test_system_users 1.37s call test_backup_restore_on_cluster/test.py::test_file_deduplication 1.34s call test_compression_nested_columns/test.py::test_nested_compression_codec 1.33s teardown test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_smaller_cluster 1.33s call test_backup_restore_on_cluster/test.py::test_different_tables_on_nodes 1.32s call test_attach_partition_with_large_destination/test.py::test_attach_partition_with_large_destination[Ordinary] 1.32s call test_attach_partition_with_large_destination/test.py::test_attach_partition_with_large_destination[Atomic] 1.30s call test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_bigger_cluster 1.28s call test_backup_restore_on_cluster/test.py::test_replicated_table_after_alters 1.23s teardown test_backup_restore_on_cluster/test.py::test_replicated_database_async 1.23s call test_backup_restore_on_cluster/test.py::test_projection 1.22s call test_concurrent_threads_soft_limit/test.py::test_use_concurrency_control_default 1.22s call test_aggregation_memory_efficient/test.py::test_remote 1.18s call test_allowed_client_hosts/test.py::test_allowed_host 1.14s call test_azure_blob_storage_zero_copy_replication/test.py::test_zero_copy_replication 1.13s call test_alternative_keeper_config/test.py::test_create_insert 1.13s teardown test_backup_restore_on_cluster/test.py::test_required_privileges 1.12s call test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[native-True] 1.08s call test_alter_codec/test.py::test_alter_codec_pk 1.06s call test_concurrent_threads_soft_limit/test.py::test_concurrent_threads_soft_limit_default 1.04s teardown test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-True] 1.03s call test_config_substitutions/test.py::test_allow_databases 1.03s teardown test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[database] 1.02s call test_backward_compatibility/test_normalized_count_comparison.py::test_select_aggregate_alias_column 1.01s call test_access_for_functions/test.py::test_access_rights_for_function 1.00s call test_backup_restore_on_cluster/test.py::test_empty_replicated_table 0.99s call test_backup_restore_on_cluster/test.py::test_table_with_parts_in_queue_considered_non_empty 0.89s teardown test_backup_restore_on_cluster/test.py::test_projection 0.88s call test_concurrent_threads_soft_limit/test.py::test_concurrent_threads_soft_limit_defined_50 0.84s call test_concurrent_threads_soft_limit/test.py::test_use_concurrency_control_soft_limit_defined_50 0.83s teardown test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_insert 0.79s call test_cluster_discovery/test_password.py::test_connect_with_password 0.78s teardown test_backup_restore_on_cluster/test.py::test_replicated_database_compare_parts 0.78s teardown test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[uuid] 0.75s call test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_smaller_cluster 0.75s call test_concurrent_threads_soft_limit/test.py::test_concurrent_threads_soft_limit_defined_1 0.74s teardown test_backup_restore_on_cluster/test.py::test_empty_replicated_table 0.73s teardown test_backup_restore_on_cluster/test.py::test_table_with_parts_in_queue_considered_non_empty 0.73s teardown test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_merge 0.71s call test_config_substitutions/test.py::test_config 0.71s teardown test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-False] 0.69s teardown test_backup_restore_on_cluster/test.py::test_keeper_value_max_size 0.69s teardown test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[native-True] 0.68s teardown test_backup_restore_on_cluster/test.py::test_mutation 0.68s teardown test_backup_restore_on_cluster/test.py::test_system_users 0.68s teardown test_backup_restore_on_cluster/test.py::test_replicated_table 0.68s teardown test_backup_restore_on_cluster/test.py::test_get_error_from_other_host 0.68s teardown test_backup_restore_on_cluster/test.py::test_table_in_replicated_database_with_not_synced_def 0.66s call test_config_substitutions/test.py::test_include_config 0.64s teardown test_backup_restore_on_cluster/test.py::test_backup_restore_on_single_replica 0.63s teardown test_backup_restore_on_cluster/test.py::test_different_tables_on_nodes 0.63s teardown test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_def 0.63s teardown test_backup_restore_on_cluster/test.py::test_shutdown_waits_for_backup 0.63s teardown test_backup_restore_on_cluster/test.py::test_system_functions 0.63s teardown test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_bigger_cluster 0.62s call test_backward_compatibility/test_select_aggregate_alias_column.py::test_select_aggregate_alias_column 0.59s teardown test_backup_restore_on_cluster/test.py::test_file_deduplication 0.59s call test_config_substitutions/test.py::test_config_from_env_overrides 0.59s call test_allowed_client_hosts/test.py::test_denied_host 0.58s teardown test_backup_restore_on_cluster/test.py::test_replicated_table_with_uuid_in_zkpath 0.58s teardown test_backup_restore_on_cluster/test.py::test_replicated_table_after_alters 0.58s teardown test_backup_restore_on_cluster/test.py::test_replicated_database 0.53s call test_buffer_profile/test.py::test_buffer_profile 0.42s call test_config_substitutions/test.py::test_config_merge_from_env_overrides 0.39s call test_attach_table_normalizer/test.py::test_attach_substr 0.39s call test_authentication/test.py::test_authentication_pass 0.38s call test_backward_compatibility/test_cte_distributed.py::test_cte_distributed 0.37s call test_custom_settings/test.py::test_custom_settings 0.31s call test_cluster_all_replicas/test.py::test_cluster 0.26s call test_backup_restore_on_cluster/test.py::test_get_error_from_other_host 0.23s call test_authentication/test.py::test_authentication_fail 0.23s call test_buffer_profile/test.py::test_default_profile 0.18s call test_cluster_all_replicas/test.py::test_global_in 0.13s call test_backward_compatibility/test_insert_profile_events.py::test_new_client_compatible 0.13s call test_compatibility_merge_tree_settings/test.py::test_config_overrides_compatibility 0.13s call test_backward_compatibility/test_insert_profile_events.py::test_old_client_compatible 0.00s setup test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[two_shards_three_nodes] 0.00s teardown test_backward_compatibility/test_functions.py::test_aggregate_states 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_merge 0.00s setup test_backup_restore_on_cluster/test.py::test_tables_dependency 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_table_after_alters 0.00s setup test_backup_restore_on_cluster/test.py::test_empty_replicated_table 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[uuid] 0.00s setup test_backup_restore_on_cluster/test.py::test_keeper_value_max_size 0.00s setup test_cgroup_limit/test.py::test_cgroup_cpu_limit 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_database 0.00s setup test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-True] 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_table 0.00s setup test_check_table/test.py::test_check_replicated_table_simple[-_0] 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_smaller_cluster 0.00s teardown test_cgroup_limit/test.py::test_cgroup_cpu_limit 0.00s setup test_attach_table_normalizer/test.py::test_attach_substr_restart 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_bigger_cluster 0.00s setup test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[native-True] 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_database_async 0.00s setup test_backup_restore_on_cluster/test.py::test_get_error_from_other_host 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_database_compare_parts 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_insert 0.00s setup test_backup_restore_on_cluster/test.py::test_shutdown_waits_for_backup 0.00s teardown test_backward_compatibility/test_insert_profile_events.py::test_new_client_compatible 0.00s setup test_backup_restore_on_cluster/test.py::test_required_privileges 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_table_with_uuid_in_zkpath 0.00s setup test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[one_shard_three_nodes] 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_def 0.00s setup test_backup_restore_on_cluster/test.py::test_table_with_parts_in_queue_considered_non_empty 0.00s setup test_backup_restore_on_cluster/test.py::test_table_in_replicated_database_with_not_synced_def 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[database] 0.00s setup test_backward_compatibility/test_functions.py::test_string_functions 0.00s setup test_backup_restore_on_cluster/test.py::test_file_deduplication 0.00s teardown test_config_substitutions/test.py::test_allow_databases 0.00s setup test_buffer_profile/test.py::test_default_profile 0.00s teardown test_attach_table_normalizer/test.py::test_attach_substr 0.00s teardown test_compatibility_merge_tree_settings/test.py::test_check_projections_compatibility 0.00s setup test_backup_restore_on_cluster/test.py::test_projection 0.00s teardown test_allowed_client_hosts/test.py::test_allowed_host 0.00s teardown test_alter_on_mixed_type_cluster/test.py::test_alter_on_cluter_non_replicated 0.00s setup test_check_table/test.py::test_check_replicated_table_corruption[-_0-.bin] 0.00s teardown test_cluster_all_replicas/test.py::test_cluster 0.00s setup test_backup_restore_on_cluster/test.py::test_system_users 0.00s teardown test_buffer_profile/test.py::test_buffer_profile 0.00s setup test_backup_restore_on_cluster/test.py::test_system_functions 0.00s setup test_compatibility_merge_tree_settings/test.py::test_config_overrides_compatibility 0.00s setup test_backward_compatibility/test_insert_profile_events.py::test_old_client_compatible 0.00s setup test_check_table/test.py::test_check_normal_table_corruption[] 0.00s teardown test_attach_partition_with_large_destination/test.py::test_attach_partition_with_large_destination[Atomic] 0.00s setup test_config_substitutions/test.py::test_config 0.00s setup test_backup_restore_on_cluster/test.py::test_mutation 0.00s setup test_backup_restore_on_cluster/test.py::test_backup_restore_on_single_replica 0.00s setup test_cluster_all_replicas/test.py::test_skip_unavailable_replica[one_shard_three_nodes] 0.00s teardown test_check_table/test.py::test_check_all_tables 0.00s teardown test_authentication/test.py::test_authentication_fail 0.00s setup test_alter_on_mixed_type_cluster/test.py::test_alter_replicated_on_cluster 0.00s setup test_backup_restore_on_cluster/test.py::test_different_tables_on_nodes 0.00s setup test_cluster_all_replicas/test.py::test_skip_unavailable_replica[two_shards_three_nodes] 0.00s setup test_attach_partition_with_large_destination/test.py::test_attach_partition_with_large_destination[Ordinary] 0.00s setup test_concurrent_threads_soft_limit/test.py::test_use_concurrency_control_soft_limit_defined_50 0.00s setup test_backup_restore/test.py::test_replace_partition 0.00s setup test_async_metrics_in_cgroup/test.py::test_system_wide_metrics 0.00s setup test_async_metrics_in_cgroup/test.py::test_user_cpu_accounting 0.00s setup test_concurrent_threads_soft_limit/test.py::test_concurrent_threads_soft_limit_defined_50 0.00s setup test_backup_restore/test.py::test_restore 0.00s setup test_authentication/test.py::test_authentication_pass 0.00s teardown test_async_metrics_in_cgroup/test.py::test_normalized_user_cpu 0.00s setup test_config_substitutions/test.py::test_config_multiple_zk_substitutions 0.00s teardown test_backup_restore/test.py::test_attach_partition 0.00s setup test_config_substitutions/test.py::test_include_config 0.00s teardown test_config_substitutions/test.py::test_config_multiple_zk_substitutions 0.00s teardown test_concurrent_threads_soft_limit/test.py::test_concurrent_threads_soft_limit_default 0.00s setup test_allowed_client_hosts/test.py::test_denied_host 0.00s setup test_config_substitutions/test.py::test_config_from_env_overrides 0.00s setup test_concurrent_threads_soft_limit/test.py::test_use_concurrency_control_default 0.00s teardown test_check_table/test.py::test_check_replicated_table_corruption[-_0-.bin] 0.00s setup test_concurrent_threads_soft_limit/test.py::test_concurrent_threads_soft_limit_defined_1 0.00s setup test_concurrent_threads_soft_limit/test.py::test_concurrent_threads_soft_limit_limit_reached 0.00s teardown test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[one_shard_three_nodes] 0.00s setup test_alter_codec/test.py::test_alter_codec_pk 0.00s teardown test_concurrent_threads_soft_limit/test.py::test_use_concurrency_control_default 0.00s teardown test_concurrent_threads_soft_limit/test.py::test_concurrent_threads_soft_limit_limit_reached 0.00s setup test_access_for_functions/test.py::test_ignore_obsolete_grant_on_database 0.00s setup test_config_substitutions/test.py::test_config_merge_from_env_overrides 0.00s setup test_cluster_all_replicas/test.py::test_global_in 0.00s teardown test_async_metrics_in_cgroup/test.py::test_system_wide_metrics 0.00s teardown test_cluster_all_replicas/test.py::test_global_in 0.00s teardown test_check_table/test.py::test_check_normal_table_corruption[] 0.00s teardown test_access_for_functions/test.py::test_access_rights_for_function 0.00s teardown test_concurrent_threads_soft_limit/test.py::test_concurrent_threads_soft_limit_defined_1 0.00s teardown test_backup_restore/test.py::test_replace_partition 0.00s teardown test_alter_codec/test.py::test_alter_codec_index 0.00s teardown test_cluster_all_replicas/test.py::test_skip_unavailable_replica[one_shard_three_nodes] 0.00s teardown test_config_substitutions/test.py::test_config_merge_from_env_overrides 0.00s teardown test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[two_shards_three_nodes] 0.00s teardown test_config_substitutions/test.py::test_config_from_env_overrides 0.00s teardown test_concurrent_threads_soft_limit/test.py::test_concurrent_threads_soft_limit_defined_50 0.00s teardown test_config_substitutions/test.py::test_config =========================== short test summary info ============================ PASSED test_alter_codec/test.py::test_alter_codec_index PASSED test_concurrent_threads_soft_limit/test.py::test_concurrent_threads_soft_limit_default PASSED test_access_for_functions/test.py::test_access_rights_for_function PASSED test_concurrent_threads_soft_limit/test.py::test_concurrent_threads_soft_limit_defined_1 PASSED test_alter_codec/test.py::test_alter_codec_pk PASSED test_concurrent_threads_soft_limit/test.py::test_concurrent_threads_soft_limit_defined_50 PASSED test_concurrent_threads_soft_limit/test.py::test_concurrent_threads_soft_limit_limit_reached PASSED test_concurrent_threads_soft_limit/test.py::test_use_concurrency_control_default PASSED test_concurrent_threads_soft_limit/test.py::test_use_concurrency_control_soft_limit_defined_50 PASSED test_access_for_functions/test.py::test_ignore_obsolete_grant_on_database PASSED test_allowed_client_hosts/test.py::test_allowed_host PASSED test_config_substitutions/test.py::test_allow_databases PASSED test_backup_restore/test.py::test_attach_partition PASSED test_allowed_client_hosts/test.py::test_denied_host PASSED test_config_substitutions/test.py::test_config PASSED test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-False] PASSED test_config_substitutions/test.py::test_config_from_env_overrides PASSED test_check_table/test.py::test_check_all_tables PASSED test_config_substitutions/test.py::test_config_merge_from_env_overrides PASSED test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-True] PASSED test_async_metrics_in_cgroup/test.py::test_normalized_user_cpu PASSED test_cluster_all_replicas/test.py::test_cluster PASSED test_backup_restore/test.py::test_replace_partition PASSED test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[native-True] PASSED test_config_substitutions/test.py::test_config_multiple_zk_substitutions PASSED test_config_substitutions/test.py::test_include_config PASSED test_backup_restore_on_cluster/test.py::test_backup_restore_on_single_replica PASSED test_backup_restore_on_cluster/test.py::test_different_tables_on_nodes PASSED test_backup_restore/test.py::test_restore PASSED test_check_table/test.py::test_check_normal_table_corruption[] PASSED test_backup_restore_on_cluster/test.py::test_empty_replicated_table PASSED test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[one_shard_three_nodes] PASSED test_authentication/test.py::test_authentication_fail PASSED test_attach_table_normalizer/test.py::test_attach_substr PASSED test_authentication/test.py::test_authentication_pass PASSED test_backup_restore_on_cluster/test.py::test_file_deduplication PASSED test_backup_restore_on_cluster/test.py::test_get_error_from_other_host PASSED test_async_metrics_in_cgroup/test.py::test_system_wide_metrics PASSED test_attach_table_normalizer/test.py::test_attach_substr_restart PASSED test_backup_restore_on_cluster/test.py::test_keeper_value_max_size PASSED test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[two_shards_three_nodes] PASSED test_cluster_all_replicas/test.py::test_global_in PASSED test_attach_partition_with_large_destination/test.py::test_attach_partition_with_large_destination[Atomic] PASSED test_alter_on_mixed_type_cluster/test.py::test_alter_on_cluter_non_replicated PASSED test_attach_partition_with_large_destination/test.py::test_attach_partition_with_large_destination[Ordinary] PASSED test_alter_on_mixed_type_cluster/test.py::test_alter_replicated_on_cluster PASSED test_backup_restore_on_cluster/test.py::test_mutation PASSED test_backup_restore_on_cluster/test.py::test_projection PASSED test_cluster_all_replicas/test.py::test_skip_unavailable_replica[one_shard_three_nodes] PASSED test_async_metrics_in_cgroup/test.py::test_user_cpu_accounting PASSED test_backup_restore_on_cluster/test.py::test_replicated_database PASSED test_alternative_keeper_config/test.py::test_create_insert PASSED test_backup_restore_on_cluster/test.py::test_replicated_database_async PASSED test_backward_compatibility/test_insert_profile_events.py::test_new_client_compatible PASSED test_backward_compatibility/test_insert_profile_events.py::test_old_client_compatible PASSED test_compatibility_merge_tree_settings/test.py::test_check_projections_compatibility PASSED test_compatibility_merge_tree_settings/test.py::test_config_overrides_compatibility PASSED test_backup_restore_on_cluster/test.py::test_replicated_database_compare_parts PASSED test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[database] PASSED test_buffer_profile/test.py::test_buffer_profile PASSED test_buffer_profile/test.py::test_default_profile PASSED test_aggregation_memory_efficient/test.py::test_remote PASSED test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[uuid] PASSED test_backup_restore_on_cluster/test.py::test_replicated_table PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_after_alters PASSED test_async_connect_to_multiple_ips/test.py::test PASSED test_backward_compatibility/test_functions.py::test_aggregate_states PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_bigger_cluster PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_smaller_cluster PASSED test_backward_compatibility/test_cte_distributed.py::test_cte_distributed PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_def PASSED test_cluster_all_replicas/test.py::test_skip_unavailable_replica[two_shards_three_nodes] PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_insert PASSED test_cluster_discovery/test_password.py::test_connect_with_password PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_merge PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_with_uuid_in_zkpath PASSED test_backup_restore_on_cluster/test.py::test_required_privileges PASSED test_build_sets_from_multiple_threads/test.py::test_set PASSED test_broken_part_during_merge/test.py::test_merge_and_part_corruption PASSED test_cleanup_after_start/test.py::test_old_dirs_cleanup PASSED test_backup_restore_on_cluster/test.py::test_shutdown_waits_for_backup PASSED test_backup_restore_on_cluster/test.py::test_system_functions PASSED test_concurrent_queries_for_all_users_restriction/test.py::test_exception_message PASSED test_backup_restore_on_cluster/test.py::test_system_users PASSED test_backup_restore_on_cluster/test.py::test_table_in_replicated_database_with_not_synced_def PASSED test_backup_restore_on_cluster/test.py::test_table_with_parts_in_queue_considered_non_empty PASSED test_backup_restore_on_cluster/test.py::test_tables_dependency PASSED test_backup_restore_new/test_shutdown_wait_backup.py::test_shutdown_wait_backup PASSED test_backward_compatibility/test_vertical_merges_from_compact_parts.py::test_vertical_merges_from_compact_parts PASSED test_backward_compatibility/test_normalized_count_comparison.py::test_select_aggregate_alias_column PASSED test_cgroup_limit/test.py::test_cgroup_cpu_limit PASSED test_cancel_freeze/test.py::test_cancel_backup PASSED test_backward_compatibility/test_convert_ordinary.py::test_convert_ordinary_to_atomic PASSED test_backward_compatibility/test_functions.py::test_string_functions PASSED test_backward_compatibility/test_select_aggregate_alias_column.py::test_select_aggregate_alias_column PASSED test_compression_nested_columns/test.py::test_nested_compression_codec PASSED test_custom_settings/test.py::test_custom_settings PASSED test_check_table/test.py::test_check_replicated_table_corruption[-_0-.bin] PASSED test_check_table/test.py::test_check_replicated_table_simple[-_0] PASSED test_azure_blob_storage_zero_copy_replication/test.py::test_zero_copy_replication ======================= 100 passed in 547.43s (0:09:07) ========================